Commit
•
b475468
1
Parent(s):
3a2f43c
add model column
Browse files- LASER2_results.csv +57 -57
- SGPT-125M-weightedmean-msmarco-specb-bitfit_results.csv +57 -57
- SGPT-125M-weightedmean-nli-bitfit_results.csv +57 -57
- SGPT-5.8B-weightedmean-msmarco-specb-bitfit_results.csv +57 -57
- SGPT-5.8B-weightedmean-nli-bitfit_results.csv +57 -57
- all-MiniLM-L6-v2_results.csv +57 -57
- all-mpnet-base-v2_results.csv +57 -57
- bert-base-uncased_results.csv +57 -57
- contriever-base-msmarco_results.csv +57 -57
- glove.6B.300d_results.csv +57 -57
- gtr-t5-base_results.csv +57 -57
- gtr-t5-xxl_results.csv +57 -57
- komninos_results.csv +57 -57
- msmarco-bert-co-condensor_results.csv +57 -57
- sentence-t5-base_results.csv +57 -57
- sentence-t5-xxl_results.csv +57 -57
- sgpt-bloom-7b1-msmarco_results.csv +57 -57
- sup-simcse-bert-base-uncased_results.csv +57 -57
- unsup-simcse-bert-base-uncased_results.csv +57 -57
LASER2_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7891044776119402
|
3 |
-
AmazonPolarityClassification,accuracy,0.6100945
|
4 |
-
AmazonReviewsClassification,accuracy,0.30578
|
5 |
-
Banking77Classification,accuracy,0.5775974025974026
|
6 |
-
EmotionClassification,accuracy,0.24830000000000002
|
7 |
-
ImdbClassification,accuracy,0.57584
|
8 |
-
MassiveIntentClassification,accuracy,0.4790181573638197
|
9 |
-
MassiveScenarioClassification,accuracy,0.558238063214526
|
10 |
-
MTOPDomainClassification,accuracy,0.7610579115367078
|
11 |
-
MTOPIntentClassification,accuracy,0.5047423620611035
|
12 |
-
ToxicConversationsClassification,accuracy,0.5404779999999999
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.4872665534804754
|
14 |
-
ArxivClusteringP2P,v_measure,0.1776823856238192
|
15 |
-
ArxivClusteringS2S,v_measure,0.1239260518556585
|
16 |
-
BiorxivClusteringP2P,v_measure,0.12399936477309108
|
17 |
-
BiorxivClusteringS2S,v_measure,0.08827421024926384
|
18 |
-
MedrxivClusteringP2P,v_measure,0.17908142247465778
|
19 |
-
MedrxivClusteringS2S,v_measure,0.1662952889451872
|
20 |
-
RedditClustering,v_measure,0.0996202584007419
|
21 |
-
RedditClusteringP2P,v_measure,0.264185580282609
|
22 |
-
StackExchangeClustering,v_measure,0.15794084761797694
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.18627980064437472
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.11378130183913912
|
25 |
-
SprintDuplicateQuestions,ap,0.6554452704075004
|
26 |
-
TwitterSemEval2015,ap,0.5956789548901481
|
27 |
-
TwitterURLCorpus,ap,0.8147393118163185
|
28 |
-
AskUbuntuDupQuestions,map,0.4898595820868522
|
29 |
-
MindSmallReranking,map,0.24788913161151171
|
30 |
-
SciDocsRR,map,0.5498656851897086
|
31 |
-
StackOverflowDupQuestions,map,0.36983704984940896
|
32 |
-
ArguAna,ndcg_at_10,0.12856
|
33 |
-
ClimateFEVER,ndcg_at_10,0.0036
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.04119
|
35 |
-
DBPedia,ndcg_at_10,0.01526
|
36 |
-
FEVER,ndcg_at_10,0.00767
|
37 |
-
FiQA2018,ndcg_at_10,0.01725
|
38 |
-
HotpotQA,ndcg_at_10,0.05504
|
39 |
-
MSMARCO,ndcg_at_10,0.03593
|
40 |
-
NFCorpus,ndcg_at_10,0.02439
|
41 |
-
NQ,ndcg_at_10,0.00642
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.71145
|
43 |
-
SCIDOCS,ndcg_at_10,0.00777
|
44 |
-
SciFact,ndcg_at_10,0.04038
|
45 |
-
Touche2020,ndcg_at_10,0.0106
|
46 |
-
TRECCOVID,ndcg_at_10,0.10972
|
47 |
-
BIOSSES,cosine_spearman,0.6200864182463187
|
48 |
-
SICK-R,cosine_spearman,0.6285761430269268
|
49 |
-
STS12,cosine_spearman,0.6260375600516247
|
50 |
-
STS13,cosine_spearman,0.5961753521490076
|
51 |
-
STS14,cosine_spearman,0.5702962555736466
|
52 |
-
STS15,cosine_spearman,0.7156560343816677
|
53 |
-
STS16,cosine_spearman,0.7075338261767046
|
54 |
-
STS17,cosine_spearman,0.7672812705124583
|
55 |
-
STS22,cosine_spearman,0.3974427076488482
|
56 |
-
STSBenchmark,cosine_spearman,0.6977241730309153
|
57 |
-
SummEval,cosine_spearman,0.268033428993562
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
LASER2,AmazonCounterfactualClassification,accuracy,0.7891044776119402
|
3 |
+
LASER2,AmazonPolarityClassification,accuracy,0.6100945
|
4 |
+
LASER2,AmazonReviewsClassification,accuracy,0.30578
|
5 |
+
LASER2,Banking77Classification,accuracy,0.5775974025974026
|
6 |
+
LASER2,EmotionClassification,accuracy,0.24830000000000002
|
7 |
+
LASER2,ImdbClassification,accuracy,0.57584
|
8 |
+
LASER2,MassiveIntentClassification,accuracy,0.4790181573638197
|
9 |
+
LASER2,MassiveScenarioClassification,accuracy,0.558238063214526
|
10 |
+
LASER2,MTOPDomainClassification,accuracy,0.7610579115367078
|
11 |
+
LASER2,MTOPIntentClassification,accuracy,0.5047423620611035
|
12 |
+
LASER2,ToxicConversationsClassification,accuracy,0.5404779999999999
|
13 |
+
LASER2,TweetSentimentExtractionClassification,accuracy,0.4872665534804754
|
14 |
+
LASER2,ArxivClusteringP2P,v_measure,0.1776823856238192
|
15 |
+
LASER2,ArxivClusteringS2S,v_measure,0.1239260518556585
|
16 |
+
LASER2,BiorxivClusteringP2P,v_measure,0.12399936477309108
|
17 |
+
LASER2,BiorxivClusteringS2S,v_measure,0.08827421024926384
|
18 |
+
LASER2,MedrxivClusteringP2P,v_measure,0.17908142247465778
|
19 |
+
LASER2,MedrxivClusteringS2S,v_measure,0.1662952889451872
|
20 |
+
LASER2,RedditClustering,v_measure,0.0996202584007419
|
21 |
+
LASER2,RedditClusteringP2P,v_measure,0.264185580282609
|
22 |
+
LASER2,StackExchangeClustering,v_measure,0.15794084761797694
|
23 |
+
LASER2,StackExchangeClusteringP2P,v_measure,0.18627980064437472
|
24 |
+
LASER2,TwentyNewsgroupsClustering,v_measure,0.11378130183913912
|
25 |
+
LASER2,SprintDuplicateQuestions,ap,0.6554452704075004
|
26 |
+
LASER2,TwitterSemEval2015,ap,0.5956789548901481
|
27 |
+
LASER2,TwitterURLCorpus,ap,0.8147393118163185
|
28 |
+
LASER2,AskUbuntuDupQuestions,map,0.4898595820868522
|
29 |
+
LASER2,MindSmallReranking,map,0.24788913161151171
|
30 |
+
LASER2,SciDocsRR,map,0.5498656851897086
|
31 |
+
LASER2,StackOverflowDupQuestions,map,0.36983704984940896
|
32 |
+
LASER2,ArguAna,ndcg_at_10,0.12856
|
33 |
+
LASER2,ClimateFEVER,ndcg_at_10,0.0036
|
34 |
+
LASER2,CQADupstackRetrieval,ndcg_at_10,0.04119
|
35 |
+
LASER2,DBPedia,ndcg_at_10,0.01526
|
36 |
+
LASER2,FEVER,ndcg_at_10,0.00767
|
37 |
+
LASER2,FiQA2018,ndcg_at_10,0.01725
|
38 |
+
LASER2,HotpotQA,ndcg_at_10,0.05504
|
39 |
+
LASER2,MSMARCO,ndcg_at_10,0.03593
|
40 |
+
LASER2,NFCorpus,ndcg_at_10,0.02439
|
41 |
+
LASER2,NQ,ndcg_at_10,0.00642
|
42 |
+
LASER2,QuoraRetrieval,ndcg_at_10,0.71145
|
43 |
+
LASER2,SCIDOCS,ndcg_at_10,0.00777
|
44 |
+
LASER2,SciFact,ndcg_at_10,0.04038
|
45 |
+
LASER2,Touche2020,ndcg_at_10,0.0106
|
46 |
+
LASER2,TRECCOVID,ndcg_at_10,0.10972
|
47 |
+
LASER2,BIOSSES,cosine_spearman,0.6200864182463187
|
48 |
+
LASER2,SICK-R,cosine_spearman,0.6285761430269268
|
49 |
+
LASER2,STS12,cosine_spearman,0.6260375600516247
|
50 |
+
LASER2,STS13,cosine_spearman,0.5961753521490076
|
51 |
+
LASER2,STS14,cosine_spearman,0.5702962555736466
|
52 |
+
LASER2,STS15,cosine_spearman,0.7156560343816677
|
53 |
+
LASER2,STS16,cosine_spearman,0.7075338261767046
|
54 |
+
LASER2,STS17,cosine_spearman,0.7672812705124583
|
55 |
+
LASER2,STS22,cosine_spearman,0.3974427076488482
|
56 |
+
LASER2,STSBenchmark,cosine_spearman,0.6977241730309153
|
57 |
+
LASER2,SummEval,cosine_spearman,0.268033428993562
|
SGPT-125M-weightedmean-msmarco-specb-bitfit_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6123880597014926
|
3 |
-
AmazonPolarityClassification,accuracy,0.65401225
|
4 |
-
AmazonReviewsClassification,accuracy,0.31165999999999994
|
5 |
-
Banking77Classification,accuracy,0.7770454545454545
|
6 |
-
EmotionClassification,accuracy,0.39075000000000004
|
7 |
-
ImdbClassification,accuracy,0.586696
|
8 |
-
MassiveIntentClassification,accuracy,0.6140887693342301
|
9 |
-
MassiveScenarioClassification,accuracy,0.6973772696704774
|
10 |
-
MTOPDomainClassification,accuracy,0.8695622435020519
|
11 |
-
MTOPIntentClassification,accuracy,0.6225262197902417
|
12 |
-
ToxicConversationsClassification,accuracy,0.6265799999999999
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5240803621958121
|
14 |
-
ArxivClusteringP2P,v_measure,0.3970858340673288
|
15 |
-
ArxivClusteringS2S,v_measure,0.2824284771372105
|
16 |
-
BiorxivClusteringP2P,v_measure,0.33632603955439844
|
17 |
-
BiorxivClusteringS2S,v_measure,0.27038042665369927
|
18 |
-
MedrxivClusteringP2P,v_measure,0.31374938993074253
|
19 |
-
MedrxivClusteringS2S,v_measure,0.26871455379644094
|
20 |
-
RedditClustering,v_measure,0.4023390747226228
|
21 |
-
RedditClusteringP2P,v_measure,0.49090518272935624
|
22 |
-
StackExchangeClustering,v_measure,0.5274481093815175
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3265999453562101
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.3212697126747911
|
25 |
-
SprintDuplicateQuestions,ap,0.8988577913120002
|
26 |
-
TwitterSemEval2015,ap,0.5474680676121269
|
27 |
-
TwitterURLCorpus,ap,0.8105760818661524
|
28 |
-
AskUbuntuDupQuestions,map,0.5583700395192394
|
29 |
-
MindSmallReranking,map,0.30402396942935334
|
30 |
-
SciDocsRR,map,0.7133941904192648
|
31 |
-
StackOverflowDupQuestions,map,0.44744984645554653
|
32 |
-
ArguAna,ndcg_at_10,0.47481
|
33 |
-
ClimateFEVER,ndcg_at_10,0.21345
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.2724866666666666
|
35 |
-
DBPedia,ndcg_at_10,0.22365
|
36 |
-
FEVER,ndcg_at_10,
|
37 |
-
FiQA2018,ndcg_at_10,0.21472
|
38 |
-
HotpotQA,ndcg_at_10,
|
39 |
-
MSMARCO,ndcg_at_10,0.54238
|
40 |
-
NFCorpus,ndcg_at_10,0.2303
|
41 |
-
NQ,ndcg_at_10,0.27858
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.82688
|
43 |
-
SCIDOCS,ndcg_at_10,0.12677
|
44 |
-
SciFact,ndcg_at_10,0.57881
|
45 |
-
Touche2020,ndcg_at_10,0.21525
|
46 |
-
TRECCOVID,ndcg_at_10,0.57921
|
47 |
-
BIOSSES,cosine_spearman,0.7520954502580506
|
48 |
-
SICK-R,cosine_spearman,0.7520954502580506
|
49 |
-
STS12,cosine_spearman,0.6652980061546658
|
50 |
-
STS13,cosine_spearman,0.7616628863742361
|
51 |
-
STS14,cosine_spearman,0.6904572664009687
|
52 |
-
STS15,cosine_spearman,0.7923677712825851
|
53 |
-
STS16,cosine_spearman,0.7606792422441928
|
54 |
-
STS17,cosine_spearman,0.8495398260629698
|
55 |
-
STS22,cosine_spearman,0.6566183708171826
|
56 |
-
STSBenchmark,cosine_spearman,0.7533716094627373
|
57 |
-
SummEval,cosine_spearman,0.2890145030911965
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,AmazonCounterfactualClassification,accuracy,0.6123880597014926
|
3 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,AmazonPolarityClassification,accuracy,0.65401225
|
4 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,AmazonReviewsClassification,accuracy,0.31165999999999994
|
5 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,Banking77Classification,accuracy,0.7770454545454545
|
6 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,EmotionClassification,accuracy,0.39075000000000004
|
7 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,ImdbClassification,accuracy,0.586696
|
8 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MassiveIntentClassification,accuracy,0.6140887693342301
|
9 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MassiveScenarioClassification,accuracy,0.6973772696704774
|
10 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MTOPDomainClassification,accuracy,0.8695622435020519
|
11 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MTOPIntentClassification,accuracy,0.6225262197902417
|
12 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,ToxicConversationsClassification,accuracy,0.6265799999999999
|
13 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,TweetSentimentExtractionClassification,accuracy,0.5240803621958121
|
14 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,ArxivClusteringP2P,v_measure,0.3970858340673288
|
15 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,ArxivClusteringS2S,v_measure,0.2824284771372105
|
16 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,BiorxivClusteringP2P,v_measure,0.33632603955439844
|
17 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,BiorxivClusteringS2S,v_measure,0.27038042665369927
|
18 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MedrxivClusteringP2P,v_measure,0.31374938993074253
|
19 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MedrxivClusteringS2S,v_measure,0.26871455379644094
|
20 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,RedditClustering,v_measure,0.4023390747226228
|
21 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,RedditClusteringP2P,v_measure,0.49090518272935624
|
22 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,StackExchangeClustering,v_measure,0.5274481093815175
|
23 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,StackExchangeClusteringP2P,v_measure,0.3265999453562101
|
24 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,TwentyNewsgroupsClustering,v_measure,0.3212697126747911
|
25 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,SprintDuplicateQuestions,ap,0.8988577913120002
|
26 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,TwitterSemEval2015,ap,0.5474680676121269
|
27 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,TwitterURLCorpus,ap,0.8105760818661524
|
28 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,AskUbuntuDupQuestions,map,0.5583700395192394
|
29 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MindSmallReranking,map,0.30402396942935334
|
30 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,SciDocsRR,map,0.7133941904192648
|
31 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,StackOverflowDupQuestions,map,0.44744984645554653
|
32 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,ArguAna,ndcg_at_10,0.47481
|
33 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,ClimateFEVER,ndcg_at_10,0.21345
|
34 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,CQADupstackRetrieval,ndcg_at_10,0.2724866666666666
|
35 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,DBPedia,ndcg_at_10,0.22365
|
36 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,FEVER,ndcg_at_10,0.55509
|
37 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,FiQA2018,ndcg_at_10,0.21472
|
38 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,HotpotQA,ndcg_at_10,0.39335
|
39 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,MSMARCO,ndcg_at_10,0.54238
|
40 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,NFCorpus,ndcg_at_10,0.2303
|
41 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,NQ,ndcg_at_10,0.27858
|
42 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,QuoraRetrieval,ndcg_at_10,0.82688
|
43 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,SCIDOCS,ndcg_at_10,0.12677
|
44 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,SciFact,ndcg_at_10,0.57881
|
45 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,Touche2020,ndcg_at_10,0.21525
|
46 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,TRECCOVID,ndcg_at_10,0.57921
|
47 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,BIOSSES,cosine_spearman,0.7520954502580506
|
48 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,SICK-R,cosine_spearman,0.7520954502580506
|
49 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS12,cosine_spearman,0.6652980061546658
|
50 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS13,cosine_spearman,0.7616628863742361
|
51 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS14,cosine_spearman,0.6904572664009687
|
52 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS15,cosine_spearman,0.7923677712825851
|
53 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS16,cosine_spearman,0.7606792422441928
|
54 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS17,cosine_spearman,0.8495398260629698
|
55 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STS22,cosine_spearman,0.6566183708171826
|
56 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,STSBenchmark,cosine_spearman,0.7533716094627373
|
57 |
+
SGPT-125M-weightedmean-msmarco-specb-bitfit,SummEval,cosine_spearman,0.2890145030911965
|
SGPT-125M-weightedmean-nli-bitfit_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6588059701492537
|
3 |
-
AmazonPolarityClassification,accuracy,0.74938225
|
4 |
-
AmazonReviewsClassification,accuracy,0.35098
|
5 |
-
Banking77Classification,accuracy,0.7467857142857142
|
6 |
-
EmotionClassification,accuracy,0.42225
|
7 |
-
ImdbClassification,accuracy,0.62896
|
8 |
-
MassiveIntentClassification,accuracy,0.5808002689979825
|
9 |
-
MassiveScenarioClassification,accuracy,0.6633826496301277
|
10 |
-
MTOPDomainClassification,accuracy,0.8151846785225718
|
11 |
-
MTOPIntentClassification,accuracy,0.582421340629275
|
12 |
-
ToxicConversationsClassification,accuracy,0.627862
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5482173174872665
|
14 |
-
ArxivClusteringP2P,v_measure,0.3474248247787077
|
15 |
-
ArxivClusteringS2S,v_measure,0.2467870651472156
|
16 |
-
BiorxivClusteringP2P,v_measure,0.2893427045246491
|
17 |
-
BiorxivClusteringS2S,v_measure,0.23080939123955474
|
18 |
-
MedrxivClusteringP2P,v_measure,0.28301902023313874
|
19 |
-
MedrxivClusteringS2S,v_measure,0.24932123582259286
|
20 |
-
RedditClustering,v_measure,0.3375741018380938
|
21 |
-
RedditClusteringP2P,v_measure,0.7765601981302583
|
22 |
-
StackExchangeClustering,v_measure,0.4459127540530939
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.28230204578753637
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.2824295128553035
|
25 |
-
SprintDuplicateQuestions,ap,0.7773085502917281
|
26 |
-
TwitterSemEval2015,ap,0.5709409536692154
|
27 |
-
TwitterURLCorpus,ap,0.805106377126291
|
28 |
-
AskUbuntuDupQuestions,map,0.5263439984994702
|
29 |
-
MindSmallReranking,map,0.2926934104146833
|
30 |
-
SciDocsRR,map,0.6835710819755543
|
31 |
-
StackOverflowDupQuestions,map,0.3996520488022785
|
32 |
-
ArguAna,ndcg_at_10,0.3104
|
33 |
-
ClimateFEVER,ndcg_at_10,0.11007
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.20290166666666667
|
35 |
-
DBPedia,ndcg_at_10,0.10868
|
36 |
-
FEVER,ndcg_at_10,0.18399
|
37 |
-
FiQA2018,ndcg_at_10,0.0894
|
38 |
-
HotpotQA,ndcg_at_10,0.17727
|
39 |
-
MSMARCO,ndcg_at_10,0.18159
|
40 |
-
NFCorpus,ndcg_at_10,0.11805
|
41 |
-
NQ,ndcg_at_10,0.07626
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.7896
|
43 |
-
SCIDOCS,ndcg_at_10,0.07133
|
44 |
-
SciFact,ndcg_at_10,0.31787
|
45 |
-
Touche2020,ndcg_at_10,0.1227
|
46 |
-
TRECCOVID,ndcg_at_10,0.3931
|
47 |
-
BIOSSES,cosine_spearman,0.7092812216947605
|
48 |
-
SICK-R,cosine_spearman,0.7092812216947605
|
49 |
-
STS12,cosine_spearman,0.6916914010333395
|
50 |
-
STS13,cosine_spearman,0.7723446219694267
|
51 |
-
STS14,cosine_spearman,0.7098892671568665
|
52 |
-
STS15,cosine_spearman,0.7973794782598049
|
53 |
-
STS16,cosine_spearman,0.7793230986462234
|
54 |
-
STS17,cosine_spearman,0.8732720847297224
|
55 |
-
STS22,cosine_spearman,0.5963676007113087
|
56 |
-
STSBenchmark,cosine_spearman,0.7954269211027273
|
57 |
-
SummEval,cosine_spearman,0.30256847004390486
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
SGPT-125M-weightedmean-nli-bitfit,AmazonCounterfactualClassification,accuracy,0.6588059701492537
|
3 |
+
SGPT-125M-weightedmean-nli-bitfit,AmazonPolarityClassification,accuracy,0.74938225
|
4 |
+
SGPT-125M-weightedmean-nli-bitfit,AmazonReviewsClassification,accuracy,0.35098
|
5 |
+
SGPT-125M-weightedmean-nli-bitfit,Banking77Classification,accuracy,0.7467857142857142
|
6 |
+
SGPT-125M-weightedmean-nli-bitfit,EmotionClassification,accuracy,0.42225
|
7 |
+
SGPT-125M-weightedmean-nli-bitfit,ImdbClassification,accuracy,0.62896
|
8 |
+
SGPT-125M-weightedmean-nli-bitfit,MassiveIntentClassification,accuracy,0.5808002689979825
|
9 |
+
SGPT-125M-weightedmean-nli-bitfit,MassiveScenarioClassification,accuracy,0.6633826496301277
|
10 |
+
SGPT-125M-weightedmean-nli-bitfit,MTOPDomainClassification,accuracy,0.8151846785225718
|
11 |
+
SGPT-125M-weightedmean-nli-bitfit,MTOPIntentClassification,accuracy,0.582421340629275
|
12 |
+
SGPT-125M-weightedmean-nli-bitfit,ToxicConversationsClassification,accuracy,0.627862
|
13 |
+
SGPT-125M-weightedmean-nli-bitfit,TweetSentimentExtractionClassification,accuracy,0.5482173174872665
|
14 |
+
SGPT-125M-weightedmean-nli-bitfit,ArxivClusteringP2P,v_measure,0.3474248247787077
|
15 |
+
SGPT-125M-weightedmean-nli-bitfit,ArxivClusteringS2S,v_measure,0.2467870651472156
|
16 |
+
SGPT-125M-weightedmean-nli-bitfit,BiorxivClusteringP2P,v_measure,0.2893427045246491
|
17 |
+
SGPT-125M-weightedmean-nli-bitfit,BiorxivClusteringS2S,v_measure,0.23080939123955474
|
18 |
+
SGPT-125M-weightedmean-nli-bitfit,MedrxivClusteringP2P,v_measure,0.28301902023313874
|
19 |
+
SGPT-125M-weightedmean-nli-bitfit,MedrxivClusteringS2S,v_measure,0.24932123582259286
|
20 |
+
SGPT-125M-weightedmean-nli-bitfit,RedditClustering,v_measure,0.3375741018380938
|
21 |
+
SGPT-125M-weightedmean-nli-bitfit,RedditClusteringP2P,v_measure,0.7765601981302583
|
22 |
+
SGPT-125M-weightedmean-nli-bitfit,StackExchangeClustering,v_measure,0.4459127540530939
|
23 |
+
SGPT-125M-weightedmean-nli-bitfit,StackExchangeClusteringP2P,v_measure,0.28230204578753637
|
24 |
+
SGPT-125M-weightedmean-nli-bitfit,TwentyNewsgroupsClustering,v_measure,0.2824295128553035
|
25 |
+
SGPT-125M-weightedmean-nli-bitfit,SprintDuplicateQuestions,ap,0.7773085502917281
|
26 |
+
SGPT-125M-weightedmean-nli-bitfit,TwitterSemEval2015,ap,0.5709409536692154
|
27 |
+
SGPT-125M-weightedmean-nli-bitfit,TwitterURLCorpus,ap,0.805106377126291
|
28 |
+
SGPT-125M-weightedmean-nli-bitfit,AskUbuntuDupQuestions,map,0.5263439984994702
|
29 |
+
SGPT-125M-weightedmean-nli-bitfit,MindSmallReranking,map,0.2926934104146833
|
30 |
+
SGPT-125M-weightedmean-nli-bitfit,SciDocsRR,map,0.6835710819755543
|
31 |
+
SGPT-125M-weightedmean-nli-bitfit,StackOverflowDupQuestions,map,0.3996520488022785
|
32 |
+
SGPT-125M-weightedmean-nli-bitfit,ArguAna,ndcg_at_10,0.3104
|
33 |
+
SGPT-125M-weightedmean-nli-bitfit,ClimateFEVER,ndcg_at_10,0.11007
|
34 |
+
SGPT-125M-weightedmean-nli-bitfit,CQADupstackRetrieval,ndcg_at_10,0.20290166666666667
|
35 |
+
SGPT-125M-weightedmean-nli-bitfit,DBPedia,ndcg_at_10,0.10868
|
36 |
+
SGPT-125M-weightedmean-nli-bitfit,FEVER,ndcg_at_10,0.18399
|
37 |
+
SGPT-125M-weightedmean-nli-bitfit,FiQA2018,ndcg_at_10,0.0894
|
38 |
+
SGPT-125M-weightedmean-nli-bitfit,HotpotQA,ndcg_at_10,0.17727
|
39 |
+
SGPT-125M-weightedmean-nli-bitfit,MSMARCO,ndcg_at_10,0.18159
|
40 |
+
SGPT-125M-weightedmean-nli-bitfit,NFCorpus,ndcg_at_10,0.11805
|
41 |
+
SGPT-125M-weightedmean-nli-bitfit,NQ,ndcg_at_10,0.07626
|
42 |
+
SGPT-125M-weightedmean-nli-bitfit,QuoraRetrieval,ndcg_at_10,0.7896
|
43 |
+
SGPT-125M-weightedmean-nli-bitfit,SCIDOCS,ndcg_at_10,0.07133
|
44 |
+
SGPT-125M-weightedmean-nli-bitfit,SciFact,ndcg_at_10,0.31787
|
45 |
+
SGPT-125M-weightedmean-nli-bitfit,Touche2020,ndcg_at_10,0.1227
|
46 |
+
SGPT-125M-weightedmean-nli-bitfit,TRECCOVID,ndcg_at_10,0.3931
|
47 |
+
SGPT-125M-weightedmean-nli-bitfit,BIOSSES,cosine_spearman,0.7092812216947605
|
48 |
+
SGPT-125M-weightedmean-nli-bitfit,SICK-R,cosine_spearman,0.7092812216947605
|
49 |
+
SGPT-125M-weightedmean-nli-bitfit,STS12,cosine_spearman,0.6916914010333395
|
50 |
+
SGPT-125M-weightedmean-nli-bitfit,STS13,cosine_spearman,0.7723446219694267
|
51 |
+
SGPT-125M-weightedmean-nli-bitfit,STS14,cosine_spearman,0.7098892671568665
|
52 |
+
SGPT-125M-weightedmean-nli-bitfit,STS15,cosine_spearman,0.7973794782598049
|
53 |
+
SGPT-125M-weightedmean-nli-bitfit,STS16,cosine_spearman,0.7793230986462234
|
54 |
+
SGPT-125M-weightedmean-nli-bitfit,STS17,cosine_spearman,0.8732720847297224
|
55 |
+
SGPT-125M-weightedmean-nli-bitfit,STS22,cosine_spearman,0.5963676007113087
|
56 |
+
SGPT-125M-weightedmean-nli-bitfit,STSBenchmark,cosine_spearman,0.7954269211027273
|
57 |
+
SGPT-125M-weightedmean-nli-bitfit,SummEval,cosine_spearman,0.30256847004390486
|
SGPT-5.8B-weightedmean-msmarco-specb-bitfit_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6922388059701492
|
3 |
-
AmazonPolarityClassification,accuracy,0.7126109999999999
|
4 |
-
AmazonReviewsClassification,accuracy,0.39191999999999994
|
5 |
-
Banking77Classification,accuracy,0.8449350649350649
|
6 |
-
EmotionClassification,accuracy,0.49655000000000005
|
7 |
-
ImdbClassification,accuracy,0.666376
|
8 |
-
MassiveIntentClassification,accuracy,0.703866845998655
|
9 |
-
MassiveScenarioClassification,accuracy,0.7627774041694687
|
10 |
-
MTOPDomainClassification,accuracy,0.9346557227542178
|
11 |
-
MTOPIntentClassification,accuracy,0.7242134062927497
|
12 |
-
ToxicConversationsClassification,accuracy,0.67709
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5685059422750425
|
14 |
-
ArxivClusteringP2P,v_measure,0.4559037428592033
|
15 |
-
ArxivClusteringS2S,v_measure,0.3886371701986363
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3655145972298938
|
17 |
-
BiorxivClusteringS2S,v_measure,0.3369901851846774
|
18 |
-
MedrxivClusteringP2P,v_measure,0.3151174592577334
|
19 |
-
MedrxivClusteringS2S,v_measure,0.28764235987575365
|
20 |
-
RedditClustering,v_measure,0.4045148482612238
|
21 |
-
RedditClusteringP2P,v_measure,0.5574949067303913
|
22 |
-
StackExchangeClustering,v_measure,0.5920812266121527
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.33954248554638056
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.3946372298609047
|
25 |
-
SprintDuplicateQuestions,ap,0.9384024096781063
|
26 |
-
TwitterSemEval2015,ap,0.6686677647503386
|
27 |
-
TwitterURLCorpus,ap,0.8528960532524223
|
28 |
-
AskUbuntuDupQuestions,map,0.6162556869142777
|
29 |
-
MindSmallReranking,map,
|
30 |
-
SciDocsRR,map,0.8078825425914723
|
31 |
-
StackOverflowDupQuestions,map,0.5152800990025549
|
32 |
-
ArguAna,ndcg_at_10,0.52203
|
33 |
-
ClimateFEVER,ndcg_at_10,
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.3940316666666666
|
35 |
-
DBPedia,ndcg_at_10,
|
36 |
-
FEVER,ndcg_at_10,
|
37 |
-
FiQA2018,ndcg_at_10,0.37463
|
38 |
-
HotpotQA,ndcg_at_10,
|
39 |
-
MSMARCO,ndcg_at_10,
|
40 |
-
NFCorpus,ndcg_at_10,0.36161
|
41 |
-
NQ,ndcg_at_10,
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.87899
|
43 |
-
SCIDOCS,ndcg_at_10,0.19899
|
44 |
-
SciFact,ndcg_at_10,0.74485
|
45 |
-
Touche2020,ndcg_at_10,
|
46 |
-
TRECCOVID,ndcg_at_10,0.77592
|
47 |
-
BIOSSES,cosine_spearman,0.862501580394454
|
48 |
-
SICK-R,cosine_spearman,0.862501580394454
|
49 |
-
STS12,cosine_spearman,0.675020551515597
|
50 |
-
STS13,cosine_spearman,0.7916333501772059
|
51 |
-
STS14,cosine_spearman,0.7446048326701329
|
52 |
-
STS15,cosine_spearman,0.8447194637929274
|
53 |
-
STS16,cosine_spearman,0.8096089203722137
|
54 |
-
STS17,cosine_spearman,0.8778202647220289
|
55 |
-
STS22,cosine_spearman,0.6934889515492327
|
56 |
-
STSBenchmark,cosine_spearman,0.8138977341532744
|
57 |
-
SummEval,cosine_spearman,0.24747448399760644
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AmazonCounterfactualClassification,accuracy,0.6922388059701492
|
3 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AmazonPolarityClassification,accuracy,0.7126109999999999
|
4 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AmazonReviewsClassification,accuracy,0.39191999999999994
|
5 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,Banking77Classification,accuracy,0.8449350649350649
|
6 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,EmotionClassification,accuracy,0.49655000000000005
|
7 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ImdbClassification,accuracy,0.666376
|
8 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MassiveIntentClassification,accuracy,0.703866845998655
|
9 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MassiveScenarioClassification,accuracy,0.7627774041694687
|
10 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MTOPDomainClassification,accuracy,0.9346557227542178
|
11 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MTOPIntentClassification,accuracy,0.7242134062927497
|
12 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ToxicConversationsClassification,accuracy,0.67709
|
13 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TweetSentimentExtractionClassification,accuracy,0.5685059422750425
|
14 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ArxivClusteringP2P,v_measure,0.4559037428592033
|
15 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ArxivClusteringS2S,v_measure,0.3886371701986363
|
16 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,BiorxivClusteringP2P,v_measure,0.3655145972298938
|
17 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,BiorxivClusteringS2S,v_measure,0.3369901851846774
|
18 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MedrxivClusteringP2P,v_measure,0.3151174592577334
|
19 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MedrxivClusteringS2S,v_measure,0.28764235987575365
|
20 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,RedditClustering,v_measure,0.4045148482612238
|
21 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,RedditClusteringP2P,v_measure,0.5574949067303913
|
22 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,StackExchangeClustering,v_measure,0.5920812266121527
|
23 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,StackExchangeClusteringP2P,v_measure,0.33954248554638056
|
24 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TwentyNewsgroupsClustering,v_measure,0.3946372298609047
|
25 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SprintDuplicateQuestions,ap,0.9384024096781063
|
26 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TwitterSemEval2015,ap,0.6686677647503386
|
27 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TwitterURLCorpus,ap,0.8528960532524223
|
28 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,AskUbuntuDupQuestions,map,0.6162556869142777
|
29 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MindSmallReranking,map,
|
30 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SciDocsRR,map,0.8078825425914723
|
31 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,StackOverflowDupQuestions,map,0.5152800990025549
|
32 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ArguAna,ndcg_at_10,0.52203
|
33 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,ClimateFEVER,ndcg_at_10,
|
34 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,CQADupstackRetrieval,ndcg_at_10,0.3940316666666666
|
35 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,DBPedia,ndcg_at_10,
|
36 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,FEVER,ndcg_at_10,
|
37 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,FiQA2018,ndcg_at_10,0.37463
|
38 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,HotpotQA,ndcg_at_10,
|
39 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,MSMARCO,ndcg_at_10,
|
40 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,NFCorpus,ndcg_at_10,0.36161
|
41 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,NQ,ndcg_at_10,
|
42 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,QuoraRetrieval,ndcg_at_10,0.87899
|
43 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SCIDOCS,ndcg_at_10,0.19899
|
44 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SciFact,ndcg_at_10,0.74485
|
45 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,Touche2020,ndcg_at_10,
|
46 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,TRECCOVID,ndcg_at_10,0.77592
|
47 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,BIOSSES,cosine_spearman,0.862501580394454
|
48 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SICK-R,cosine_spearman,0.862501580394454
|
49 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS12,cosine_spearman,0.675020551515597
|
50 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS13,cosine_spearman,0.7916333501772059
|
51 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS14,cosine_spearman,0.7446048326701329
|
52 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS15,cosine_spearman,0.8447194637929274
|
53 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS16,cosine_spearman,0.8096089203722137
|
54 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS17,cosine_spearman,0.8778202647220289
|
55 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STS22,cosine_spearman,0.6934889515492327
|
56 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,STSBenchmark,cosine_spearman,0.8138977341532744
|
57 |
+
SGPT-5.8B-weightedmean-msmarco-specb-bitfit,SummEval,cosine_spearman,0.24747448399760644
|
SGPT-5.8B-weightedmean-nli-bitfit_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7407462686567164
|
3 |
-
AmazonPolarityClassification,accuracy,0.8230920000000002
|
4 |
-
AmazonReviewsClassification,accuracy,0.41584000000000004
|
5 |
-
Banking77Classification,accuracy,0.8174025974025974
|
6 |
-
EmotionClassification,accuracy,0.4991999999999999
|
7 |
-
ImdbClassification,accuracy,0.7432920000000001
|
8 |
-
MassiveIntentClassification,accuracy,0.7000336247478144
|
9 |
-
MassiveScenarioClassification,accuracy,0.7502689979825151
|
10 |
-
MTOPDomainClassification,accuracy,0.896374829001368
|
11 |
-
MTOPIntentClassification,accuracy,0.7068171454628362
|
12 |
-
ToxicConversationsClassification,accuracy,0.6993339999999999
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.6243916242218449
|
14 |
-
ArxivClusteringP2P,v_measure,0.40553923271901693
|
15 |
-
ArxivClusteringS2S,v_measure,0.3249323183712211
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3359451202614059
|
17 |
-
BiorxivClusteringS2S,v_measure,0.29128241446157166
|
18 |
-
MedrxivClusteringP2P,v_measure,0.30327566856300814
|
19 |
-
MedrxivClusteringS2S,v_measure,0.2801650210863619
|
20 |
-
RedditClustering,v_measure,0.42171313610410677
|
21 |
-
RedditClusteringP2P,v_measure,0.4801815621479994
|
22 |
-
StackExchangeClustering,v_measure,0.5413314692311623
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.31115181648287144
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.37202082549859794
|
25 |
-
SprintDuplicateQuestions,ap,0.8053523657342799
|
26 |
-
TwitterSemEval2015,ap,0.6599787692764192
|
27 |
-
TwitterURLCorpus,ap,0.8454288910204419
|
28 |
-
AskUbuntuDupQuestions,map,0.5589811361443445
|
29 |
-
MindSmallReranking,map,
|
30 |
-
SciDocsRR,map,0.7753549990038017
|
31 |
-
StackOverflowDupQuestions,map,0.44771112666694335
|
32 |
-
ArguAna,ndcg_at_10,0.3507
|
33 |
-
ClimateFEVER,ndcg_at_10,0.17574
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.29978333333333335
|
35 |
-
DBPedia,ndcg_at_10,0.26098
|
36 |
-
FEVER,ndcg_at_10,0.38644
|
37 |
-
FiQA2018,ndcg_at_10,0.18591
|
38 |
-
HotpotQA,ndcg_at_10,0.33995
|
39 |
-
MSMARCO,ndcg_at_10,
|
40 |
-
NFCorpus,ndcg_at_10,0.28256
|
41 |
-
NQ,ndcg_at_10,0.24628
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.8468
|
43 |
-
SCIDOCS,ndcg_at_10,0.13554
|
44 |
-
SciFact,ndcg_at_10,0.46662
|
45 |
-
Touche2020,ndcg_at_10,0.16178
|
46 |
-
TRECCOVID,ndcg_at_10,0.55353
|
47 |
-
BIOSSES,cosine_spearman,0.7950000423261177
|
48 |
-
SICK-R,cosine_spearman,0.7958562467776268
|
49 |
-
STS12,cosine_spearman,0.7428502198729446
|
50 |
-
STS13,cosine_spearman,0.8535148434923192
|
51 |
-
STS14,cosine_spearman,0.7920612995350826
|
52 |
-
STS15,cosine_spearman,0.8552171407629341
|
53 |
-
STS16,cosine_spearman,0.8253853238521991
|
54 |
-
STS17,cosine_spearman,0.904389369935707
|
55 |
-
STS22,cosine_spearman,0.6319719780439462
|
56 |
-
STSBenchmark,cosine_spearman,0.8566816193002651
|
57 |
-
SummEval,cosine_spearman,0.30384175038360195
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
SGPT-5.8B-weightedmean-nli-bitfit,AmazonCounterfactualClassification,accuracy,0.7407462686567164
|
3 |
+
SGPT-5.8B-weightedmean-nli-bitfit,AmazonPolarityClassification,accuracy,0.8230920000000002
|
4 |
+
SGPT-5.8B-weightedmean-nli-bitfit,AmazonReviewsClassification,accuracy,0.41584000000000004
|
5 |
+
SGPT-5.8B-weightedmean-nli-bitfit,Banking77Classification,accuracy,0.8174025974025974
|
6 |
+
SGPT-5.8B-weightedmean-nli-bitfit,EmotionClassification,accuracy,0.4991999999999999
|
7 |
+
SGPT-5.8B-weightedmean-nli-bitfit,ImdbClassification,accuracy,0.7432920000000001
|
8 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MassiveIntentClassification,accuracy,0.7000336247478144
|
9 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MassiveScenarioClassification,accuracy,0.7502689979825151
|
10 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MTOPDomainClassification,accuracy,0.896374829001368
|
11 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MTOPIntentClassification,accuracy,0.7068171454628362
|
12 |
+
SGPT-5.8B-weightedmean-nli-bitfit,ToxicConversationsClassification,accuracy,0.6993339999999999
|
13 |
+
SGPT-5.8B-weightedmean-nli-bitfit,TweetSentimentExtractionClassification,accuracy,0.6243916242218449
|
14 |
+
SGPT-5.8B-weightedmean-nli-bitfit,ArxivClusteringP2P,v_measure,0.40553923271901693
|
15 |
+
SGPT-5.8B-weightedmean-nli-bitfit,ArxivClusteringS2S,v_measure,0.3249323183712211
|
16 |
+
SGPT-5.8B-weightedmean-nli-bitfit,BiorxivClusteringP2P,v_measure,0.3359451202614059
|
17 |
+
SGPT-5.8B-weightedmean-nli-bitfit,BiorxivClusteringS2S,v_measure,0.29128241446157166
|
18 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MedrxivClusteringP2P,v_measure,0.30327566856300814
|
19 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MedrxivClusteringS2S,v_measure,0.2801650210863619
|
20 |
+
SGPT-5.8B-weightedmean-nli-bitfit,RedditClustering,v_measure,0.42171313610410677
|
21 |
+
SGPT-5.8B-weightedmean-nli-bitfit,RedditClusteringP2P,v_measure,0.4801815621479994
|
22 |
+
SGPT-5.8B-weightedmean-nli-bitfit,StackExchangeClustering,v_measure,0.5413314692311623
|
23 |
+
SGPT-5.8B-weightedmean-nli-bitfit,StackExchangeClusteringP2P,v_measure,0.31115181648287144
|
24 |
+
SGPT-5.8B-weightedmean-nli-bitfit,TwentyNewsgroupsClustering,v_measure,0.37202082549859794
|
25 |
+
SGPT-5.8B-weightedmean-nli-bitfit,SprintDuplicateQuestions,ap,0.8053523657342799
|
26 |
+
SGPT-5.8B-weightedmean-nli-bitfit,TwitterSemEval2015,ap,0.6599787692764192
|
27 |
+
SGPT-5.8B-weightedmean-nli-bitfit,TwitterURLCorpus,ap,0.8454288910204419
|
28 |
+
SGPT-5.8B-weightedmean-nli-bitfit,AskUbuntuDupQuestions,map,0.5589811361443445
|
29 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MindSmallReranking,map,
|
30 |
+
SGPT-5.8B-weightedmean-nli-bitfit,SciDocsRR,map,0.7753549990038017
|
31 |
+
SGPT-5.8B-weightedmean-nli-bitfit,StackOverflowDupQuestions,map,0.44771112666694335
|
32 |
+
SGPT-5.8B-weightedmean-nli-bitfit,ArguAna,ndcg_at_10,0.3507
|
33 |
+
SGPT-5.8B-weightedmean-nli-bitfit,ClimateFEVER,ndcg_at_10,0.17574
|
34 |
+
SGPT-5.8B-weightedmean-nli-bitfit,CQADupstackRetrieval,ndcg_at_10,0.29978333333333335
|
35 |
+
SGPT-5.8B-weightedmean-nli-bitfit,DBPedia,ndcg_at_10,0.26098
|
36 |
+
SGPT-5.8B-weightedmean-nli-bitfit,FEVER,ndcg_at_10,0.38644
|
37 |
+
SGPT-5.8B-weightedmean-nli-bitfit,FiQA2018,ndcg_at_10,0.18591
|
38 |
+
SGPT-5.8B-weightedmean-nli-bitfit,HotpotQA,ndcg_at_10,0.33995
|
39 |
+
SGPT-5.8B-weightedmean-nli-bitfit,MSMARCO,ndcg_at_10,
|
40 |
+
SGPT-5.8B-weightedmean-nli-bitfit,NFCorpus,ndcg_at_10,0.28256
|
41 |
+
SGPT-5.8B-weightedmean-nli-bitfit,NQ,ndcg_at_10,0.24628
|
42 |
+
SGPT-5.8B-weightedmean-nli-bitfit,QuoraRetrieval,ndcg_at_10,0.8468
|
43 |
+
SGPT-5.8B-weightedmean-nli-bitfit,SCIDOCS,ndcg_at_10,0.13554
|
44 |
+
SGPT-5.8B-weightedmean-nli-bitfit,SciFact,ndcg_at_10,0.46662
|
45 |
+
SGPT-5.8B-weightedmean-nli-bitfit,Touche2020,ndcg_at_10,0.16178
|
46 |
+
SGPT-5.8B-weightedmean-nli-bitfit,TRECCOVID,ndcg_at_10,0.55353
|
47 |
+
SGPT-5.8B-weightedmean-nli-bitfit,BIOSSES,cosine_spearman,0.7950000423261177
|
48 |
+
SGPT-5.8B-weightedmean-nli-bitfit,SICK-R,cosine_spearman,0.7958562467776268
|
49 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS12,cosine_spearman,0.7428502198729446
|
50 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS13,cosine_spearman,0.8535148434923192
|
51 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS14,cosine_spearman,0.7920612995350826
|
52 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS15,cosine_spearman,0.8552171407629341
|
53 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS16,cosine_spearman,0.8253853238521991
|
54 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS17,cosine_spearman,0.904389369935707
|
55 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STS22,cosine_spearman,0.6319719780439462
|
56 |
+
SGPT-5.8B-weightedmean-nli-bitfit,STSBenchmark,cosine_spearman,0.8566816193002651
|
57 |
+
SGPT-5.8B-weightedmean-nli-bitfit,SummEval,cosine_spearman,0.30384175038360195
|
all-MiniLM-L6-v2_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.641492537313433
|
3 |
-
AmazonPolarityClassification,accuracy,0.62582975
|
4 |
-
AmazonReviewsClassification,accuracy,0.31786
|
5 |
-
Banking77Classification,accuracy,0.7975000000000001
|
6 |
-
EmotionClassification,accuracy,0.3843
|
7 |
-
ImdbClassification,accuracy,0.606612
|
8 |
-
MassiveIntentClassification,accuracy,0.6740416946872899
|
9 |
-
MassiveScenarioClassification,accuracy,0.7575655682582381
|
10 |
-
MTOPDomainClassification,accuracy,0.9156178750569997
|
11 |
-
MTOPIntentClassification,accuracy,0.6218194254445966
|
12 |
-
ToxicConversationsClassification,accuracy,0.669918
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.554102999434069
|
14 |
-
ArxivClusteringP2P,v_measure,0.46545950790501556
|
15 |
-
ArxivClusteringS2S,v_measure,0.3785709823840442
|
16 |
-
BiorxivClusteringP2P,v_measure,0.38483019141351227
|
17 |
-
BiorxivClusteringS2S,v_measure,0.33170209943399803
|
18 |
-
MedrxivClusteringP2P,v_measure,0.3440873490143895
|
19 |
-
MedrxivClusteringS2S,v_measure,0.3229220750053091
|
20 |
-
RedditClustering,v_measure,0.5066969274980475
|
21 |
-
RedditClusteringP2P,v_measure,0.8150529020581544
|
22 |
-
StackExchangeClustering,v_measure,0.5336142166203601
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3800182562780098
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.4686027142764777
|
25 |
-
SprintDuplicateQuestions,ap,0.9455063045792446
|
26 |
-
TwitterSemEval2015,ap,0.6785802440228593
|
27 |
-
TwitterURLCorpus,ap,0.847016595145111
|
28 |
-
AskUbuntuDupQuestions,map,0.634768168123733
|
29 |
-
MindSmallReranking,map,0.30798042020200267
|
30 |
-
SciDocsRR,map,0.8711941318470207
|
31 |
-
StackOverflowDupQuestions,map,0.5076213438431608
|
32 |
-
ArguAna,ndcg_at_10,0.50167
|
33 |
-
ClimateFEVER,ndcg_at_10,0.20272
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.4132475
|
35 |
-
DBPedia,ndcg_at_10,0.3233
|
36 |
-
FEVER,ndcg_at_10,0.51934
|
37 |
-
FiQA2018,ndcg_at_10,0.36867
|
38 |
-
HotpotQA,ndcg_at_10,0.46513
|
39 |
-
MSMARCO,ndcg_at_10,0.63685
|
40 |
-
NFCorpus,ndcg_at_10,0.31594
|
41 |
-
NQ,ndcg_at_10,0.43869
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.87556
|
43 |
-
SCIDOCS,ndcg_at_10,0.21641
|
44 |
-
SciFact,ndcg_at_10,0.64508
|
45 |
-
Touche2020,ndcg_at_10,0.16904
|
46 |
-
TRECCOVID,ndcg_at_10,0.47246
|
47 |
-
BIOSSES,cosine_spearman,0.8164041444909368
|
48 |
-
SICK-R,cosine_spearman,0.7758245130495687
|
49 |
-
STS12,cosine_spearman,0.7236900735029991
|
50 |
-
STS13,cosine_spearman,0.8060316722220763
|
51 |
-
STS14,cosine_spearman,0.7558912800301661
|
52 |
-
STS15,cosine_spearman,0.8538966051883823
|
53 |
-
STS16,cosine_spearman,0.7898945343973262
|
54 |
-
STS17,cosine_spearman,0.8758779089494525
|
55 |
-
STS22,cosine_spearman,0.6721465212910986
|
56 |
-
STSBenchmark,cosine_spearman,0.8203246731235654
|
57 |
-
SummEval,cosine_spearman,0.3080875767124449
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
all-MiniLM-L6-v2,AmazonCounterfactualClassification,accuracy,0.641492537313433
|
3 |
+
all-MiniLM-L6-v2,AmazonPolarityClassification,accuracy,0.62582975
|
4 |
+
all-MiniLM-L6-v2,AmazonReviewsClassification,accuracy,0.31786
|
5 |
+
all-MiniLM-L6-v2,Banking77Classification,accuracy,0.7975000000000001
|
6 |
+
all-MiniLM-L6-v2,EmotionClassification,accuracy,0.3843
|
7 |
+
all-MiniLM-L6-v2,ImdbClassification,accuracy,0.606612
|
8 |
+
all-MiniLM-L6-v2,MassiveIntentClassification,accuracy,0.6740416946872899
|
9 |
+
all-MiniLM-L6-v2,MassiveScenarioClassification,accuracy,0.7575655682582381
|
10 |
+
all-MiniLM-L6-v2,MTOPDomainClassification,accuracy,0.9156178750569997
|
11 |
+
all-MiniLM-L6-v2,MTOPIntentClassification,accuracy,0.6218194254445966
|
12 |
+
all-MiniLM-L6-v2,ToxicConversationsClassification,accuracy,0.669918
|
13 |
+
all-MiniLM-L6-v2,TweetSentimentExtractionClassification,accuracy,0.554102999434069
|
14 |
+
all-MiniLM-L6-v2,ArxivClusteringP2P,v_measure,0.46545950790501556
|
15 |
+
all-MiniLM-L6-v2,ArxivClusteringS2S,v_measure,0.3785709823840442
|
16 |
+
all-MiniLM-L6-v2,BiorxivClusteringP2P,v_measure,0.38483019141351227
|
17 |
+
all-MiniLM-L6-v2,BiorxivClusteringS2S,v_measure,0.33170209943399803
|
18 |
+
all-MiniLM-L6-v2,MedrxivClusteringP2P,v_measure,0.3440873490143895
|
19 |
+
all-MiniLM-L6-v2,MedrxivClusteringS2S,v_measure,0.3229220750053091
|
20 |
+
all-MiniLM-L6-v2,RedditClustering,v_measure,0.5066969274980475
|
21 |
+
all-MiniLM-L6-v2,RedditClusteringP2P,v_measure,0.8150529020581544
|
22 |
+
all-MiniLM-L6-v2,StackExchangeClustering,v_measure,0.5336142166203601
|
23 |
+
all-MiniLM-L6-v2,StackExchangeClusteringP2P,v_measure,0.3800182562780098
|
24 |
+
all-MiniLM-L6-v2,TwentyNewsgroupsClustering,v_measure,0.4686027142764777
|
25 |
+
all-MiniLM-L6-v2,SprintDuplicateQuestions,ap,0.9455063045792446
|
26 |
+
all-MiniLM-L6-v2,TwitterSemEval2015,ap,0.6785802440228593
|
27 |
+
all-MiniLM-L6-v2,TwitterURLCorpus,ap,0.847016595145111
|
28 |
+
all-MiniLM-L6-v2,AskUbuntuDupQuestions,map,0.634768168123733
|
29 |
+
all-MiniLM-L6-v2,MindSmallReranking,map,0.30798042020200267
|
30 |
+
all-MiniLM-L6-v2,SciDocsRR,map,0.8711941318470207
|
31 |
+
all-MiniLM-L6-v2,StackOverflowDupQuestions,map,0.5076213438431608
|
32 |
+
all-MiniLM-L6-v2,ArguAna,ndcg_at_10,0.50167
|
33 |
+
all-MiniLM-L6-v2,ClimateFEVER,ndcg_at_10,0.20272
|
34 |
+
all-MiniLM-L6-v2,CQADupstackRetrieval,ndcg_at_10,0.4132475
|
35 |
+
all-MiniLM-L6-v2,DBPedia,ndcg_at_10,0.3233
|
36 |
+
all-MiniLM-L6-v2,FEVER,ndcg_at_10,0.51934
|
37 |
+
all-MiniLM-L6-v2,FiQA2018,ndcg_at_10,0.36867
|
38 |
+
all-MiniLM-L6-v2,HotpotQA,ndcg_at_10,0.46513
|
39 |
+
all-MiniLM-L6-v2,MSMARCO,ndcg_at_10,0.63685
|
40 |
+
all-MiniLM-L6-v2,NFCorpus,ndcg_at_10,0.31594
|
41 |
+
all-MiniLM-L6-v2,NQ,ndcg_at_10,0.43869
|
42 |
+
all-MiniLM-L6-v2,QuoraRetrieval,ndcg_at_10,0.87556
|
43 |
+
all-MiniLM-L6-v2,SCIDOCS,ndcg_at_10,0.21641
|
44 |
+
all-MiniLM-L6-v2,SciFact,ndcg_at_10,0.64508
|
45 |
+
all-MiniLM-L6-v2,Touche2020,ndcg_at_10,0.16904
|
46 |
+
all-MiniLM-L6-v2,TRECCOVID,ndcg_at_10,0.47246
|
47 |
+
all-MiniLM-L6-v2,BIOSSES,cosine_spearman,0.8164041444909368
|
48 |
+
all-MiniLM-L6-v2,SICK-R,cosine_spearman,0.7758245130495687
|
49 |
+
all-MiniLM-L6-v2,STS12,cosine_spearman,0.7236900735029991
|
50 |
+
all-MiniLM-L6-v2,STS13,cosine_spearman,0.8060316722220763
|
51 |
+
all-MiniLM-L6-v2,STS14,cosine_spearman,0.7558912800301661
|
52 |
+
all-MiniLM-L6-v2,STS15,cosine_spearman,0.8538966051883823
|
53 |
+
all-MiniLM-L6-v2,STS16,cosine_spearman,0.7898945343973262
|
54 |
+
all-MiniLM-L6-v2,STS17,cosine_spearman,0.8758779089494525
|
55 |
+
all-MiniLM-L6-v2,STS22,cosine_spearman,0.6721465212910986
|
56 |
+
all-MiniLM-L6-v2,STSBenchmark,cosine_spearman,0.8203246731235654
|
57 |
+
all-MiniLM-L6-v2,SummEval,cosine_spearman,0.3080875767124449
|
all-mpnet-base-v2_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.652686567164179
|
3 |
-
AmazonPolarityClassification,accuracy,0.6713145
|
4 |
-
AmazonReviewsClassification,accuracy,0.31920000000000004
|
5 |
-
Banking77Classification,accuracy,0.8186038961038961
|
6 |
-
EmotionClassification,accuracy,0.39725
|
7 |
-
ImdbClassification,accuracy,0.707156
|
8 |
-
MassiveIntentClassification,accuracy,0.6956624075319435
|
9 |
-
MassiveScenarioClassification,accuracy,0.7601210490921319
|
10 |
-
MTOPDomainClassification,accuracy,0.9207934336525309
|
11 |
-
MTOPIntentClassification,accuracy,0.7020975832193344
|
12 |
-
ToxicConversationsClassification,accuracy,0.608612
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5546406338426711
|
14 |
-
ArxivClusteringP2P,v_measure,0.48378863409223744
|
15 |
-
ArxivClusteringS2S,v_measure,0.3972488615315985
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3961660513362518
|
17 |
-
BiorxivClusteringS2S,v_measure,0.3502442407186902
|
18 |
-
MedrxivClusteringP2P,v_measure,0.3558002813186373
|
19 |
-
MedrxivClusteringS2S,v_measure,0.32872725562410443
|
20 |
-
RedditClustering,v_measure,0.5482425669843732
|
21 |
-
RedditClusteringP2P,v_measure,0.5676897267804937
|
22 |
-
StackExchangeClustering,v_measure,0.5380048409076928
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.34280269334397545
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.49744553480838094
|
25 |
-
SprintDuplicateQuestions,ap,0.9014551416831161
|
26 |
-
TwitterSemEval2015,ap,0.7385398650568216
|
27 |
-
TwitterURLCorpus,ap,0.8510831084479727
|
28 |
-
AskUbuntuDupQuestions,map,0.6585199009344481
|
29 |
-
MindSmallReranking,map,0.3096534360486133
|
30 |
-
SciDocsRR,map,0.8865396986895777
|
31 |
-
StackOverflowDupQuestions,map,0.5197907654945493
|
32 |
-
ArguAna,ndcg_at_10,0.46521
|
33 |
-
ClimateFEVER,ndcg_at_10,0.21971
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.44959166666666667
|
35 |
-
DBPedia,ndcg_at_10,0.32088
|
36 |
-
FEVER,ndcg_at_10,0.50863
|
37 |
-
FiQA2018,ndcg_at_10,0.49963
|
38 |
-
HotpotQA,ndcg_at_10,0.39294
|
39 |
-
MSMARCO,ndcg_at_10,0.66682
|
40 |
-
NFCorpus,ndcg_at_10,0.33289
|
41 |
-
NQ,ndcg_at_10,0.5045
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.87462
|
43 |
-
SCIDOCS,ndcg_at_10,0.23765
|
44 |
-
SciFact,ndcg_at_10,0.6557
|
45 |
-
Touche2020,ndcg_at_10,0.19934
|
46 |
-
TRECCOVID,ndcg_at_10,0.51326
|
47 |
-
BIOSSES,cosine_spearman,0.8043195317854409
|
48 |
-
SICK-R,cosine_spearman,0.8058541928440089
|
49 |
-
STS12,cosine_spearman,0.726342785382245
|
50 |
-
STS13,cosine_spearman,0.8348499016384896
|
51 |
-
STS14,cosine_spearman,0.7800013713421681
|
52 |
-
STS15,cosine_spearman,0.856627539164948
|
53 |
-
STS16,cosine_spearman,0.8003011315645663
|
54 |
-
STS17,cosine_spearman,0.9059523263123733
|
55 |
-
STS22,cosine_spearman,0.6794534221542502
|
56 |
-
STSBenchmark,cosine_spearman,0.83421898505618
|
57 |
-
SummEval,cosine_spearman,0.27485869639926136
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
all-mpnet-base-v2,AmazonCounterfactualClassification,accuracy,0.652686567164179
|
3 |
+
all-mpnet-base-v2,AmazonPolarityClassification,accuracy,0.6713145
|
4 |
+
all-mpnet-base-v2,AmazonReviewsClassification,accuracy,0.31920000000000004
|
5 |
+
all-mpnet-base-v2,Banking77Classification,accuracy,0.8186038961038961
|
6 |
+
all-mpnet-base-v2,EmotionClassification,accuracy,0.39725
|
7 |
+
all-mpnet-base-v2,ImdbClassification,accuracy,0.707156
|
8 |
+
all-mpnet-base-v2,MassiveIntentClassification,accuracy,0.6956624075319435
|
9 |
+
all-mpnet-base-v2,MassiveScenarioClassification,accuracy,0.7601210490921319
|
10 |
+
all-mpnet-base-v2,MTOPDomainClassification,accuracy,0.9207934336525309
|
11 |
+
all-mpnet-base-v2,MTOPIntentClassification,accuracy,0.7020975832193344
|
12 |
+
all-mpnet-base-v2,ToxicConversationsClassification,accuracy,0.608612
|
13 |
+
all-mpnet-base-v2,TweetSentimentExtractionClassification,accuracy,0.5546406338426711
|
14 |
+
all-mpnet-base-v2,ArxivClusteringP2P,v_measure,0.48378863409223744
|
15 |
+
all-mpnet-base-v2,ArxivClusteringS2S,v_measure,0.3972488615315985
|
16 |
+
all-mpnet-base-v2,BiorxivClusteringP2P,v_measure,0.3961660513362518
|
17 |
+
all-mpnet-base-v2,BiorxivClusteringS2S,v_measure,0.3502442407186902
|
18 |
+
all-mpnet-base-v2,MedrxivClusteringP2P,v_measure,0.3558002813186373
|
19 |
+
all-mpnet-base-v2,MedrxivClusteringS2S,v_measure,0.32872725562410443
|
20 |
+
all-mpnet-base-v2,RedditClustering,v_measure,0.5482425669843732
|
21 |
+
all-mpnet-base-v2,RedditClusteringP2P,v_measure,0.5676897267804937
|
22 |
+
all-mpnet-base-v2,StackExchangeClustering,v_measure,0.5380048409076928
|
23 |
+
all-mpnet-base-v2,StackExchangeClusteringP2P,v_measure,0.34280269334397545
|
24 |
+
all-mpnet-base-v2,TwentyNewsgroupsClustering,v_measure,0.49744553480838094
|
25 |
+
all-mpnet-base-v2,SprintDuplicateQuestions,ap,0.9014551416831161
|
26 |
+
all-mpnet-base-v2,TwitterSemEval2015,ap,0.7385398650568216
|
27 |
+
all-mpnet-base-v2,TwitterURLCorpus,ap,0.8510831084479727
|
28 |
+
all-mpnet-base-v2,AskUbuntuDupQuestions,map,0.6585199009344481
|
29 |
+
all-mpnet-base-v2,MindSmallReranking,map,0.3096534360486133
|
30 |
+
all-mpnet-base-v2,SciDocsRR,map,0.8865396986895777
|
31 |
+
all-mpnet-base-v2,StackOverflowDupQuestions,map,0.5197907654945493
|
32 |
+
all-mpnet-base-v2,ArguAna,ndcg_at_10,0.46521
|
33 |
+
all-mpnet-base-v2,ClimateFEVER,ndcg_at_10,0.21971
|
34 |
+
all-mpnet-base-v2,CQADupstackRetrieval,ndcg_at_10,0.44959166666666667
|
35 |
+
all-mpnet-base-v2,DBPedia,ndcg_at_10,0.32088
|
36 |
+
all-mpnet-base-v2,FEVER,ndcg_at_10,0.50863
|
37 |
+
all-mpnet-base-v2,FiQA2018,ndcg_at_10,0.49963
|
38 |
+
all-mpnet-base-v2,HotpotQA,ndcg_at_10,0.39294
|
39 |
+
all-mpnet-base-v2,MSMARCO,ndcg_at_10,0.66682
|
40 |
+
all-mpnet-base-v2,NFCorpus,ndcg_at_10,0.33289
|
41 |
+
all-mpnet-base-v2,NQ,ndcg_at_10,0.5045
|
42 |
+
all-mpnet-base-v2,QuoraRetrieval,ndcg_at_10,0.87462
|
43 |
+
all-mpnet-base-v2,SCIDOCS,ndcg_at_10,0.23765
|
44 |
+
all-mpnet-base-v2,SciFact,ndcg_at_10,0.6557
|
45 |
+
all-mpnet-base-v2,Touche2020,ndcg_at_10,0.19934
|
46 |
+
all-mpnet-base-v2,TRECCOVID,ndcg_at_10,0.51326
|
47 |
+
all-mpnet-base-v2,BIOSSES,cosine_spearman,0.8043195317854409
|
48 |
+
all-mpnet-base-v2,SICK-R,cosine_spearman,0.8058541928440089
|
49 |
+
all-mpnet-base-v2,STS12,cosine_spearman,0.726342785382245
|
50 |
+
all-mpnet-base-v2,STS13,cosine_spearman,0.8348499016384896
|
51 |
+
all-mpnet-base-v2,STS14,cosine_spearman,0.7800013713421681
|
52 |
+
all-mpnet-base-v2,STS15,cosine_spearman,0.856627539164948
|
53 |
+
all-mpnet-base-v2,STS16,cosine_spearman,0.8003011315645663
|
54 |
+
all-mpnet-base-v2,STS17,cosine_spearman,0.9059523263123733
|
55 |
+
all-mpnet-base-v2,STS22,cosine_spearman,0.6794534221542502
|
56 |
+
all-mpnet-base-v2,STSBenchmark,cosine_spearman,0.83421898505618
|
57 |
+
all-mpnet-base-v2,SummEval,cosine_spearman,0.27485869639926136
|
bert-base-uncased_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7425373134328358
|
3 |
-
AmazonPolarityClassification,accuracy,0.7132945
|
4 |
-
AmazonReviewsClassification,accuracy,0.33564
|
5 |
-
Banking77Classification,accuracy,0.6340584415584415
|
6 |
-
EmotionClassification,accuracy,0.3528
|
7 |
-
ImdbClassification,accuracy,0.653456
|
8 |
-
MassiveIntentClassification,accuracy,0.5988231338264962
|
9 |
-
MassiveScenarioClassification,accuracy,0.6427706792199059
|
10 |
-
MTOPDomainClassification,accuracy,0.8262653898768809
|
11 |
-
MTOPIntentClassification,accuracy,0.6813725490196079
|
12 |
-
ToxicConversationsClassification,accuracy,0.6999679999999999
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5180814940577249
|
14 |
-
ArxivClusteringP2P,v_measure,0.3518932830729758
|
15 |
-
ArxivClusteringS2S,v_measure,0.2750822951507033
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3012278640553615
|
17 |
-
BiorxivClusteringS2S,v_measure,0.2476601053047804
|
18 |
-
MedrxivClusteringP2P,v_measure,0.26087688306606044
|
19 |
-
MedrxivClusteringS2S,v_measure,0.23604914608225602
|
20 |
-
RedditClustering,v_measure,0.27241718557705
|
21 |
-
RedditClusteringP2P,v_measure,0.43323854460464056
|
22 |
-
StackExchangeClustering,v_measure,0.4358261870547655
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.2654851675795123
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.23354321403189832
|
25 |
-
SprintDuplicateQuestions,ap,0.36808703728970593
|
26 |
-
TwitterSemEval2015,ap,0.5589788752976392
|
27 |
-
TwitterURLCorpus,ap,0.7628732783437441
|
28 |
-
AskUbuntuDupQuestions,map,0.4584088706528255
|
29 |
-
MindSmallReranking,map,0.28366637355845425
|
30 |
-
SciDocsRR,map,0.6493728987830145
|
31 |
-
StackOverflowDupQuestions,map,0.34615478798860166
|
32 |
-
ArguAna,ndcg_at_10,0.28294
|
33 |
-
ClimateFEVER,ndcg_at_10,0.0541
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
|
35 |
-
DBPedia,ndcg_at_10,0.04132
|
36 |
-
FEVER,ndcg_at_10,0.033
|
37 |
-
FiQA2018,ndcg_at_10,0.02191
|
38 |
-
HotpotQA,ndcg_at_10,0.0826
|
39 |
-
MSMARCO,ndcg_at_10,0.06176
|
40 |
-
NFCorpus,ndcg_at_10,0.04304
|
41 |
-
NQ,ndcg_at_10,0.02615
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.61029
|
43 |
-
SCIDOCS,ndcg_at_10,0.02815
|
44 |
-
SciFact,ndcg_at_10,0.13339
|
45 |
-
Touche2020,ndcg_at_10,0.00967
|
46 |
-
TRECCOVID,ndcg_at_10,0.14745
|
47 |
-
BIOSSES,cosine_spearman,0.5469823428818151
|
48 |
-
SICK-R,cosine_spearman,0.5864506948179484
|
49 |
-
STS12,cosine_spearman,0.30871788407575457
|
50 |
-
STS13,cosine_spearman,0.5989485045425808
|
51 |
-
STS14,cosine_spearman,0.4772791352844941
|
52 |
-
STS15,cosine_spearman,0.6028567383446486
|
53 |
-
STS16,cosine_spearman,0.6373272693604403
|
54 |
-
STS17,cosine_spearman,0.6410023907260639
|
55 |
-
STS22,cosine_spearman,0.563667996535454
|
56 |
-
STSBenchmark,cosine_spearman,0.4729108172174081
|
57 |
-
SummEval,cosine_spearman,0.2981716355664815
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7425373134328358
|
3 |
+
bert-base-uncased,AmazonPolarityClassification,accuracy,0.7132945
|
4 |
+
bert-base-uncased,AmazonReviewsClassification,accuracy,0.33564
|
5 |
+
bert-base-uncased,Banking77Classification,accuracy,0.6340584415584415
|
6 |
+
bert-base-uncased,EmotionClassification,accuracy,0.3528
|
7 |
+
bert-base-uncased,ImdbClassification,accuracy,0.653456
|
8 |
+
bert-base-uncased,MassiveIntentClassification,accuracy,0.5988231338264962
|
9 |
+
bert-base-uncased,MassiveScenarioClassification,accuracy,0.6427706792199059
|
10 |
+
bert-base-uncased,MTOPDomainClassification,accuracy,0.8262653898768809
|
11 |
+
bert-base-uncased,MTOPIntentClassification,accuracy,0.6813725490196079
|
12 |
+
bert-base-uncased,ToxicConversationsClassification,accuracy,0.6999679999999999
|
13 |
+
bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5180814940577249
|
14 |
+
bert-base-uncased,ArxivClusteringP2P,v_measure,0.3518932830729758
|
15 |
+
bert-base-uncased,ArxivClusteringS2S,v_measure,0.2750822951507033
|
16 |
+
bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3012278640553615
|
17 |
+
bert-base-uncased,BiorxivClusteringS2S,v_measure,0.2476601053047804
|
18 |
+
bert-base-uncased,MedrxivClusteringP2P,v_measure,0.26087688306606044
|
19 |
+
bert-base-uncased,MedrxivClusteringS2S,v_measure,0.23604914608225602
|
20 |
+
bert-base-uncased,RedditClustering,v_measure,0.27241718557705
|
21 |
+
bert-base-uncased,RedditClusteringP2P,v_measure,0.43323854460464056
|
22 |
+
bert-base-uncased,StackExchangeClustering,v_measure,0.4358261870547655
|
23 |
+
bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2654851675795123
|
24 |
+
bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23354321403189832
|
25 |
+
bert-base-uncased,SprintDuplicateQuestions,ap,0.36808703728970593
|
26 |
+
bert-base-uncased,TwitterSemEval2015,ap,0.5589788752976392
|
27 |
+
bert-base-uncased,TwitterURLCorpus,ap,0.7628732783437441
|
28 |
+
bert-base-uncased,AskUbuntuDupQuestions,map,0.4584088706528255
|
29 |
+
bert-base-uncased,MindSmallReranking,map,0.28366637355845425
|
30 |
+
bert-base-uncased,SciDocsRR,map,0.6493728987830145
|
31 |
+
bert-base-uncased,StackOverflowDupQuestions,map,0.34615478798860166
|
32 |
+
bert-base-uncased,ArguAna,ndcg_at_10,0.28294
|
33 |
+
bert-base-uncased,ClimateFEVER,ndcg_at_10,0.0541
|
34 |
+
bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
|
35 |
+
bert-base-uncased,DBPedia,ndcg_at_10,0.04132
|
36 |
+
bert-base-uncased,FEVER,ndcg_at_10,0.033
|
37 |
+
bert-base-uncased,FiQA2018,ndcg_at_10,0.02191
|
38 |
+
bert-base-uncased,HotpotQA,ndcg_at_10,0.0826
|
39 |
+
bert-base-uncased,MSMARCO,ndcg_at_10,0.06176
|
40 |
+
bert-base-uncased,NFCorpus,ndcg_at_10,0.04304
|
41 |
+
bert-base-uncased,NQ,ndcg_at_10,0.02615
|
42 |
+
bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.61029
|
43 |
+
bert-base-uncased,SCIDOCS,ndcg_at_10,0.02815
|
44 |
+
bert-base-uncased,SciFact,ndcg_at_10,0.13339
|
45 |
+
bert-base-uncased,Touche2020,ndcg_at_10,0.00967
|
46 |
+
bert-base-uncased,TRECCOVID,ndcg_at_10,0.14745
|
47 |
+
bert-base-uncased,BIOSSES,cosine_spearman,0.5469823428818151
|
48 |
+
bert-base-uncased,SICK-R,cosine_spearman,0.5864506948179484
|
49 |
+
bert-base-uncased,STS12,cosine_spearman,0.30871788407575457
|
50 |
+
bert-base-uncased,STS13,cosine_spearman,0.5989485045425808
|
51 |
+
bert-base-uncased,STS14,cosine_spearman,0.4772791352844941
|
52 |
+
bert-base-uncased,STS15,cosine_spearman,0.6028567383446486
|
53 |
+
bert-base-uncased,STS16,cosine_spearman,0.6373272693604403
|
54 |
+
bert-base-uncased,STS17,cosine_spearman,0.6410023907260639
|
55 |
+
bert-base-uncased,STS22,cosine_spearman,0.563667996535454
|
56 |
+
bert-base-uncased,STSBenchmark,cosine_spearman,0.4729108172174081
|
57 |
+
bert-base-uncased,SummEval,cosine_spearman,0.2981716355664815
|
contriever-base-msmarco_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7219402985074627
|
3 |
-
AmazonPolarityClassification,accuracy,0.6863262499999999
|
4 |
-
AmazonReviewsClassification,accuracy,0.37422
|
5 |
-
Banking77Classification,accuracy,0.8001623376623377
|
6 |
-
EmotionClassification,accuracy,0.44770000000000004
|
7 |
-
ImdbClassification,accuracy,0.670388
|
8 |
-
MassiveIntentClassification,accuracy,0.6777740416946872
|
9 |
-
MassiveScenarioClassification,accuracy,0.7599865501008743
|
10 |
-
MTOPDomainClassification,accuracy,0.9317829457364342
|
11 |
-
MTOPIntentClassification,accuracy,0.6931372549019608
|
12 |
-
ToxicConversationsClassification,accuracy,0.6776540000000001
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5610356536502547
|
14 |
-
ArxivClusteringP2P,v_measure,0.42609650289899825
|
15 |
-
ArxivClusteringS2S,v_measure,0.32317533856057756
|
16 |
-
BiorxivClusteringP2P,v_measure,0.34974621074365847
|
17 |
-
BiorxivClusteringS2S,v_measure,0.29083876572182377
|
18 |
-
MedrxivClusteringP2P,v_measure,0.3118675108904548
|
19 |
-
MedrxivClusteringS2S,v_measure,0.2727180592101203
|
20 |
-
RedditClustering,v_measure,0.5489383973987073
|
21 |
-
RedditClusteringP2P,v_measure,0.5757965400787036
|
22 |
-
StackExchangeClustering,v_measure,0.6314901527677189
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3225478381926452
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.46822120585974714
|
25 |
-
SprintDuplicateQuestions,ap,0.9554507015913917
|
26 |
-
TwitterSemEval2015,ap,0.6684972822513366
|
27 |
-
TwitterURLCorpus,ap,0.8520869965638479
|
28 |
-
AskUbuntuDupQuestions,map,0.5668809109301288
|
29 |
-
MindSmallReranking,map,0.3157849515126634
|
30 |
-
SciDocsRR,map,0.7650730255518788
|
31 |
-
StackOverflowDupQuestions,map,0.4778138013068855
|
32 |
-
ArguAna,ndcg_at_10,0.4832
|
33 |
-
ClimateFEVER,ndcg_at_10,0.24788
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.3366891666666667
|
35 |
-
DBPedia,ndcg_at_10,0.38103
|
36 |
-
FEVER,ndcg_at_10,0.59294
|
37 |
-
FiQA2018,ndcg_at_10,0.2742
|
38 |
-
HotpotQA,ndcg_at_10,0.56808
|
39 |
-
MSMARCO,ndcg_at_10,0.62494
|
40 |
-
NFCorpus,ndcg_at_10,0.31315
|
41 |
-
NQ,ndcg_at_10,0.41833
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.86719
|
43 |
-
SCIDOCS,ndcg_at_10,0.17125
|
44 |
-
SciFact,ndcg_at_10,0.6551
|
45 |
-
Touche2020,ndcg_at_10,0.15789
|
46 |
-
TRECCOVID,ndcg_at_10,0.44769
|
47 |
-
BIOSSES,cosine_spearman,0.8331816352111794
|
48 |
-
SICK-R,cosine_spearman,0.702003740199912
|
49 |
-
STS12,cosine_spearman,0.643399816144592
|
50 |
-
STS13,cosine_spearman,0.8003351945178154
|
51 |
-
STS14,cosine_spearman,0.7450891052510561
|
52 |
-
STS15,cosine_spearman,0.8329896673390621
|
53 |
-
STS16,cosine_spearman,0.7967145670824619
|
54 |
-
STS17,cosine_spearman,0.863170912160602
|
55 |
-
STS22,cosine_spearman,0.6463838183557634
|
56 |
-
STSBenchmark,cosine_spearman,0.7880678352766406
|
57 |
-
SummEval,cosine_spearman,0.30361250428863945
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
contriever-base-msmarco,AmazonCounterfactualClassification,accuracy,0.7219402985074627
|
3 |
+
contriever-base-msmarco,AmazonPolarityClassification,accuracy,0.6863262499999999
|
4 |
+
contriever-base-msmarco,AmazonReviewsClassification,accuracy,0.37422
|
5 |
+
contriever-base-msmarco,Banking77Classification,accuracy,0.8001623376623377
|
6 |
+
contriever-base-msmarco,EmotionClassification,accuracy,0.44770000000000004
|
7 |
+
contriever-base-msmarco,ImdbClassification,accuracy,0.670388
|
8 |
+
contriever-base-msmarco,MassiveIntentClassification,accuracy,0.6777740416946872
|
9 |
+
contriever-base-msmarco,MassiveScenarioClassification,accuracy,0.7599865501008743
|
10 |
+
contriever-base-msmarco,MTOPDomainClassification,accuracy,0.9317829457364342
|
11 |
+
contriever-base-msmarco,MTOPIntentClassification,accuracy,0.6931372549019608
|
12 |
+
contriever-base-msmarco,ToxicConversationsClassification,accuracy,0.6776540000000001
|
13 |
+
contriever-base-msmarco,TweetSentimentExtractionClassification,accuracy,0.5610356536502547
|
14 |
+
contriever-base-msmarco,ArxivClusteringP2P,v_measure,0.42609650289899825
|
15 |
+
contriever-base-msmarco,ArxivClusteringS2S,v_measure,0.32317533856057756
|
16 |
+
contriever-base-msmarco,BiorxivClusteringP2P,v_measure,0.34974621074365847
|
17 |
+
contriever-base-msmarco,BiorxivClusteringS2S,v_measure,0.29083876572182377
|
18 |
+
contriever-base-msmarco,MedrxivClusteringP2P,v_measure,0.3118675108904548
|
19 |
+
contriever-base-msmarco,MedrxivClusteringS2S,v_measure,0.2727180592101203
|
20 |
+
contriever-base-msmarco,RedditClustering,v_measure,0.5489383973987073
|
21 |
+
contriever-base-msmarco,RedditClusteringP2P,v_measure,0.5757965400787036
|
22 |
+
contriever-base-msmarco,StackExchangeClustering,v_measure,0.6314901527677189
|
23 |
+
contriever-base-msmarco,StackExchangeClusteringP2P,v_measure,0.3225478381926452
|
24 |
+
contriever-base-msmarco,TwentyNewsgroupsClustering,v_measure,0.46822120585974714
|
25 |
+
contriever-base-msmarco,SprintDuplicateQuestions,ap,0.9554507015913917
|
26 |
+
contriever-base-msmarco,TwitterSemEval2015,ap,0.6684972822513366
|
27 |
+
contriever-base-msmarco,TwitterURLCorpus,ap,0.8520869965638479
|
28 |
+
contriever-base-msmarco,AskUbuntuDupQuestions,map,0.5668809109301288
|
29 |
+
contriever-base-msmarco,MindSmallReranking,map,0.3157849515126634
|
30 |
+
contriever-base-msmarco,SciDocsRR,map,0.7650730255518788
|
31 |
+
contriever-base-msmarco,StackOverflowDupQuestions,map,0.4778138013068855
|
32 |
+
contriever-base-msmarco,ArguAna,ndcg_at_10,0.4832
|
33 |
+
contriever-base-msmarco,ClimateFEVER,ndcg_at_10,0.24788
|
34 |
+
contriever-base-msmarco,CQADupstackRetrieval,ndcg_at_10,0.3366891666666667
|
35 |
+
contriever-base-msmarco,DBPedia,ndcg_at_10,0.38103
|
36 |
+
contriever-base-msmarco,FEVER,ndcg_at_10,0.59294
|
37 |
+
contriever-base-msmarco,FiQA2018,ndcg_at_10,0.2742
|
38 |
+
contriever-base-msmarco,HotpotQA,ndcg_at_10,0.56808
|
39 |
+
contriever-base-msmarco,MSMARCO,ndcg_at_10,0.62494
|
40 |
+
contriever-base-msmarco,NFCorpus,ndcg_at_10,0.31315
|
41 |
+
contriever-base-msmarco,NQ,ndcg_at_10,0.41833
|
42 |
+
contriever-base-msmarco,QuoraRetrieval,ndcg_at_10,0.86719
|
43 |
+
contriever-base-msmarco,SCIDOCS,ndcg_at_10,0.17125
|
44 |
+
contriever-base-msmarco,SciFact,ndcg_at_10,0.6551
|
45 |
+
contriever-base-msmarco,Touche2020,ndcg_at_10,0.15789
|
46 |
+
contriever-base-msmarco,TRECCOVID,ndcg_at_10,0.44769
|
47 |
+
contriever-base-msmarco,BIOSSES,cosine_spearman,0.8331816352111794
|
48 |
+
contriever-base-msmarco,SICK-R,cosine_spearman,0.702003740199912
|
49 |
+
contriever-base-msmarco,STS12,cosine_spearman,0.643399816144592
|
50 |
+
contriever-base-msmarco,STS13,cosine_spearman,0.8003351945178154
|
51 |
+
contriever-base-msmarco,STS14,cosine_spearman,0.7450891052510561
|
52 |
+
contriever-base-msmarco,STS15,cosine_spearman,0.8329896673390621
|
53 |
+
contriever-base-msmarco,STS16,cosine_spearman,0.7967145670824619
|
54 |
+
contriever-base-msmarco,STS17,cosine_spearman,0.863170912160602
|
55 |
+
contriever-base-msmarco,STS22,cosine_spearman,0.6463838183557634
|
56 |
+
contriever-base-msmarco,STSBenchmark,cosine_spearman,0.7880678352766406
|
57 |
+
contriever-base-msmarco,SummEval,cosine_spearman,0.30361250428863945
|
glove.6B.300d_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.5691044776119403
|
3 |
-
AmazonPolarityClassification,accuracy,0.6032235
|
4 |
-
AmazonReviewsClassification,accuracy,0.2967000000000001
|
5 |
-
Banking77Classification,accuracy,0.6768831168831169
|
6 |
-
EmotionClassification,accuracy,0.36929999999999996
|
7 |
-
ImdbClassification,accuracy,0.625748
|
8 |
-
MassiveIntentClassification,accuracy,0.5618695359784802
|
9 |
-
MassiveScenarioClassification,accuracy,0.6602555480833894
|
10 |
-
MTOPDomainClassification,accuracy,0.791062471500228
|
11 |
-
MTOPIntentClassification,accuracy,0.5585499316005472
|
12 |
-
ToxicConversationsClassification,accuracy,0.654038
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5079513299377476
|
14 |
-
ArxivClusteringP2P,v_measure,0.3255904912728055
|
15 |
-
ArxivClusteringS2S,v_measure,0.23143703465994314
|
16 |
-
BiorxivClusteringP2P,v_measure,0.2926940233440429
|
17 |
-
BiorxivClusteringS2S,v_measure,0.19184208889576307
|
18 |
-
MedrxivClusteringP2P,v_measure,0.2612166047215606
|
19 |
-
MedrxivClusteringS2S,v_measure,0.20375267261606717
|
20 |
-
RedditClustering,v_measure,0.2845571223825329
|
21 |
-
RedditClusteringP2P,v_measure,0.7826896969479853
|
22 |
-
StackExchangeClustering,v_measure,0.35797310020384626
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.28510988741557386
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.25831825117696844
|
25 |
-
SprintDuplicateQuestions,ap,0.8696278773063988
|
26 |
-
TwitterSemEval2015,ap,0.48454642053467223
|
27 |
-
TwitterURLCorpus,ap,0.7734926843948666
|
28 |
-
AskUbuntuDupQuestions,map,0.4957211733829243
|
29 |
-
MindSmallReranking,map,0.27006016291330964
|
30 |
-
SciDocsRR,map,0.6255895801677928
|
31 |
-
StackOverflowDupQuestions,map,0.3403090098227517
|
32 |
-
ArguAna,ndcg_at_10,0.363
|
33 |
-
ClimateFEVER,ndcg_at_10,0.14441
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.154745
|
35 |
-
DBPedia,ndcg_at_10,0.18285
|
36 |
-
FEVER,ndcg_at_10,0.1499
|
37 |
-
FiQA2018,ndcg_at_10,0.10087
|
38 |
-
HotpotQA,ndcg_at_10,0.1918
|
39 |
-
MSMARCO,ndcg_at_10,0.27094
|
40 |
-
NFCorpus,ndcg_at_10,0.13871
|
41 |
-
NQ,ndcg_at_10,0.1287
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.71318
|
43 |
-
SCIDOCS,ndcg_at_10,0.08041
|
44 |
-
SciFact,ndcg_at_10,0.29575
|
45 |
-
Touche2020,ndcg_at_10,0.13987
|
46 |
-
TRECCOVID,ndcg_at_10,0.36219
|
47 |
-
BIOSSES,cosine_spearman,0.44927529728339055
|
48 |
-
SICK-R,cosine_spearman,0.5542938456737263
|
49 |
-
STS12,cosine_spearman,0.5463983809943868
|
50 |
-
STS13,cosine_spearman,0.691628709341984
|
51 |
-
STS14,cosine_spearman,0.6081333831404667
|
52 |
-
STS15,cosine_spearman,0.723122589323431
|
53 |
-
STS16,cosine_spearman,0.6533668138063549
|
54 |
-
STS17,cosine_spearman,0.7794897670754327
|
55 |
-
STS22,cosine_spearman,0.5635046917979596
|
56 |
-
STSBenchmark,cosine_spearman,0.6153556159874409
|
57 |
-
SummEval,cosine_spearman,0.2886781899399968
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
glove.6B.300d,AmazonCounterfactualClassification,accuracy,0.5691044776119403
|
3 |
+
glove.6B.300d,AmazonPolarityClassification,accuracy,0.6032235
|
4 |
+
glove.6B.300d,AmazonReviewsClassification,accuracy,0.2967000000000001
|
5 |
+
glove.6B.300d,Banking77Classification,accuracy,0.6768831168831169
|
6 |
+
glove.6B.300d,EmotionClassification,accuracy,0.36929999999999996
|
7 |
+
glove.6B.300d,ImdbClassification,accuracy,0.625748
|
8 |
+
glove.6B.300d,MassiveIntentClassification,accuracy,0.5618695359784802
|
9 |
+
glove.6B.300d,MassiveScenarioClassification,accuracy,0.6602555480833894
|
10 |
+
glove.6B.300d,MTOPDomainClassification,accuracy,0.791062471500228
|
11 |
+
glove.6B.300d,MTOPIntentClassification,accuracy,0.5585499316005472
|
12 |
+
glove.6B.300d,ToxicConversationsClassification,accuracy,0.654038
|
13 |
+
glove.6B.300d,TweetSentimentExtractionClassification,accuracy,0.5079513299377476
|
14 |
+
glove.6B.300d,ArxivClusteringP2P,v_measure,0.3255904912728055
|
15 |
+
glove.6B.300d,ArxivClusteringS2S,v_measure,0.23143703465994314
|
16 |
+
glove.6B.300d,BiorxivClusteringP2P,v_measure,0.2926940233440429
|
17 |
+
glove.6B.300d,BiorxivClusteringS2S,v_measure,0.19184208889576307
|
18 |
+
glove.6B.300d,MedrxivClusteringP2P,v_measure,0.2612166047215606
|
19 |
+
glove.6B.300d,MedrxivClusteringS2S,v_measure,0.20375267261606717
|
20 |
+
glove.6B.300d,RedditClustering,v_measure,0.2845571223825329
|
21 |
+
glove.6B.300d,RedditClusteringP2P,v_measure,0.7826896969479853
|
22 |
+
glove.6B.300d,StackExchangeClustering,v_measure,0.35797310020384626
|
23 |
+
glove.6B.300d,StackExchangeClusteringP2P,v_measure,0.28510988741557386
|
24 |
+
glove.6B.300d,TwentyNewsgroupsClustering,v_measure,0.25831825117696844
|
25 |
+
glove.6B.300d,SprintDuplicateQuestions,ap,0.8696278773063988
|
26 |
+
glove.6B.300d,TwitterSemEval2015,ap,0.48454642053467223
|
27 |
+
glove.6B.300d,TwitterURLCorpus,ap,0.7734926843948666
|
28 |
+
glove.6B.300d,AskUbuntuDupQuestions,map,0.4957211733829243
|
29 |
+
glove.6B.300d,MindSmallReranking,map,0.27006016291330964
|
30 |
+
glove.6B.300d,SciDocsRR,map,0.6255895801677928
|
31 |
+
glove.6B.300d,StackOverflowDupQuestions,map,0.3403090098227517
|
32 |
+
glove.6B.300d,ArguAna,ndcg_at_10,0.363
|
33 |
+
glove.6B.300d,ClimateFEVER,ndcg_at_10,0.14441
|
34 |
+
glove.6B.300d,CQADupstackRetrieval,ndcg_at_10,0.154745
|
35 |
+
glove.6B.300d,DBPedia,ndcg_at_10,0.18285
|
36 |
+
glove.6B.300d,FEVER,ndcg_at_10,0.1499
|
37 |
+
glove.6B.300d,FiQA2018,ndcg_at_10,0.10087
|
38 |
+
glove.6B.300d,HotpotQA,ndcg_at_10,0.1918
|
39 |
+
glove.6B.300d,MSMARCO,ndcg_at_10,0.27094
|
40 |
+
glove.6B.300d,NFCorpus,ndcg_at_10,0.13871
|
41 |
+
glove.6B.300d,NQ,ndcg_at_10,0.1287
|
42 |
+
glove.6B.300d,QuoraRetrieval,ndcg_at_10,0.71318
|
43 |
+
glove.6B.300d,SCIDOCS,ndcg_at_10,0.08041
|
44 |
+
glove.6B.300d,SciFact,ndcg_at_10,0.29575
|
45 |
+
glove.6B.300d,Touche2020,ndcg_at_10,0.13987
|
46 |
+
glove.6B.300d,TRECCOVID,ndcg_at_10,0.36219
|
47 |
+
glove.6B.300d,BIOSSES,cosine_spearman,0.44927529728339055
|
48 |
+
glove.6B.300d,SICK-R,cosine_spearman,0.5542938456737263
|
49 |
+
glove.6B.300d,STS12,cosine_spearman,0.5463983809943868
|
50 |
+
glove.6B.300d,STS13,cosine_spearman,0.691628709341984
|
51 |
+
glove.6B.300d,STS14,cosine_spearman,0.6081333831404667
|
52 |
+
glove.6B.300d,STS15,cosine_spearman,0.723122589323431
|
53 |
+
glove.6B.300d,STS16,cosine_spearman,0.6533668138063549
|
54 |
+
glove.6B.300d,STS17,cosine_spearman,0.7794897670754327
|
55 |
+
glove.6B.300d,STS22,cosine_spearman,0.5635046917979596
|
56 |
+
glove.6B.300d,STSBenchmark,cosine_spearman,0.6153556159874409
|
57 |
+
glove.6B.300d,SummEval,cosine_spearman,0.2886781899399968
|
gtr-t5-base_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6932835820895522
|
3 |
-
AmazonPolarityClassification,accuracy,0.67818775
|
4 |
-
AmazonReviewsClassification,accuracy,0.38482
|
5 |
-
Banking77Classification,accuracy,0.7925649350649351
|
6 |
-
EmotionClassification,accuracy,0.422
|
7 |
-
ImdbClassification,accuracy,0.659944
|
8 |
-
MassiveIntentClassification,accuracy,0.6705110961667786
|
9 |
-
MassiveScenarioClassification,accuracy,0.7539677202420982
|
10 |
-
MTOPDomainClassification,accuracy,0.9242134062927496
|
11 |
-
MTOPIntentClassification,accuracy,0.6244186046511628
|
12 |
-
ToxicConversationsClassification,accuracy,0.6660360000000001
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5602150537634408
|
14 |
-
ArxivClusteringP2P,v_measure,0.3548695236674728
|
15 |
-
ArxivClusteringS2S,v_measure,0.27180108855677926
|
16 |
-
BiorxivClusteringP2P,v_measure,0.2765822038298308
|
17 |
-
BiorxivClusteringS2S,v_measure,0.23251707197080881
|
18 |
-
MedrxivClusteringP2P,v_measure,0.2756514819237031
|
19 |
-
MedrxivClusteringS2S,v_measure,0.2513156119229876
|
20 |
-
RedditClustering,v_measure,0.5613269289874916
|
21 |
-
RedditClusteringP2P,v_measure,0.585344949362798
|
22 |
-
StackExchangeClustering,v_measure,0.6421338133178497
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3301453796958954
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.46718954435107624
|
25 |
-
SprintDuplicateQuestions,ap,0.9454933848900439
|
26 |
-
TwitterSemEval2015,ap,0.7222675077030647
|
27 |
-
TwitterURLCorpus,ap,0.8477048332956029
|
28 |
-
AskUbuntuDupQuestions,map,0.6086144157716742
|
29 |
-
MindSmallReranking,map,0.313303339508014
|
30 |
-
SciDocsRR,map,0.737089144684744
|
31 |
-
StackOverflowDupQuestions,map,0.510120398337891
|
32 |
-
ArguAna,ndcg_at_10,0.50828
|
33 |
-
ClimateFEVER,ndcg_at_10,0.24884
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.34548749999999995
|
35 |
-
DBPedia,ndcg_at_10,0.35244
|
36 |
-
FEVER,ndcg_at_10,0.68932
|
37 |
-
FiQA2018,ndcg_at_10,0.35147
|
38 |
-
HotpotQA,ndcg_at_10,0.54926
|
39 |
-
MSMARCO,ndcg_at_10,0.68745
|
40 |
-
NFCorpus,ndcg_at_10,0.30223
|
41 |
-
NQ,ndcg_at_10,0.50469
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.87977
|
43 |
-
SCIDOCS,ndcg_at_10,0.14
|
44 |
-
SciFact,ndcg_at_10,0.59738
|
45 |
-
Touche2020,ndcg_at_10,0.25891
|
46 |
-
TRECCOVID,ndcg_at_10,0.56047
|
47 |
-
BIOSSES,cosine_spearman,0.7900420980306923
|
48 |
-
SICK-R,cosine_spearman,0.7900420980306923
|
49 |
-
STS12,cosine_spearman,0.6859358272758398
|
50 |
-
STS13,cosine_spearman,0.7909019447848391
|
51 |
-
STS14,cosine_spearman,0.7464141703710193
|
52 |
-
STS15,cosine_spearman,0.8484757279716068
|
53 |
-
STS16,cosine_spearman,0.8156700264541153
|
54 |
-
STS17,cosine_spearman,0.8579615701004483
|
55 |
-
STS22,cosine_spearman,0.6616950180914625
|
56 |
-
STSBenchmark,cosine_spearman,0.7957572102026641
|
57 |
-
SummEval,cosine_spearman,0.29669058153671907
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
gtr-t5-base,AmazonCounterfactualClassification,accuracy,0.6932835820895522
|
3 |
+
gtr-t5-base,AmazonPolarityClassification,accuracy,0.67818775
|
4 |
+
gtr-t5-base,AmazonReviewsClassification,accuracy,0.38482
|
5 |
+
gtr-t5-base,Banking77Classification,accuracy,0.7925649350649351
|
6 |
+
gtr-t5-base,EmotionClassification,accuracy,0.422
|
7 |
+
gtr-t5-base,ImdbClassification,accuracy,0.659944
|
8 |
+
gtr-t5-base,MassiveIntentClassification,accuracy,0.6705110961667786
|
9 |
+
gtr-t5-base,MassiveScenarioClassification,accuracy,0.7539677202420982
|
10 |
+
gtr-t5-base,MTOPDomainClassification,accuracy,0.9242134062927496
|
11 |
+
gtr-t5-base,MTOPIntentClassification,accuracy,0.6244186046511628
|
12 |
+
gtr-t5-base,ToxicConversationsClassification,accuracy,0.6660360000000001
|
13 |
+
gtr-t5-base,TweetSentimentExtractionClassification,accuracy,0.5602150537634408
|
14 |
+
gtr-t5-base,ArxivClusteringP2P,v_measure,0.3548695236674728
|
15 |
+
gtr-t5-base,ArxivClusteringS2S,v_measure,0.27180108855677926
|
16 |
+
gtr-t5-base,BiorxivClusteringP2P,v_measure,0.2765822038298308
|
17 |
+
gtr-t5-base,BiorxivClusteringS2S,v_measure,0.23251707197080881
|
18 |
+
gtr-t5-base,MedrxivClusteringP2P,v_measure,0.2756514819237031
|
19 |
+
gtr-t5-base,MedrxivClusteringS2S,v_measure,0.2513156119229876
|
20 |
+
gtr-t5-base,RedditClustering,v_measure,0.5613269289874916
|
21 |
+
gtr-t5-base,RedditClusteringP2P,v_measure,0.585344949362798
|
22 |
+
gtr-t5-base,StackExchangeClustering,v_measure,0.6421338133178497
|
23 |
+
gtr-t5-base,StackExchangeClusteringP2P,v_measure,0.3301453796958954
|
24 |
+
gtr-t5-base,TwentyNewsgroupsClustering,v_measure,0.46718954435107624
|
25 |
+
gtr-t5-base,SprintDuplicateQuestions,ap,0.9454933848900439
|
26 |
+
gtr-t5-base,TwitterSemEval2015,ap,0.7222675077030647
|
27 |
+
gtr-t5-base,TwitterURLCorpus,ap,0.8477048332956029
|
28 |
+
gtr-t5-base,AskUbuntuDupQuestions,map,0.6086144157716742
|
29 |
+
gtr-t5-base,MindSmallReranking,map,0.313303339508014
|
30 |
+
gtr-t5-base,SciDocsRR,map,0.737089144684744
|
31 |
+
gtr-t5-base,StackOverflowDupQuestions,map,0.510120398337891
|
32 |
+
gtr-t5-base,ArguAna,ndcg_at_10,0.50828
|
33 |
+
gtr-t5-base,ClimateFEVER,ndcg_at_10,0.24884
|
34 |
+
gtr-t5-base,CQADupstackRetrieval,ndcg_at_10,0.34548749999999995
|
35 |
+
gtr-t5-base,DBPedia,ndcg_at_10,0.35244
|
36 |
+
gtr-t5-base,FEVER,ndcg_at_10,0.68932
|
37 |
+
gtr-t5-base,FiQA2018,ndcg_at_10,0.35147
|
38 |
+
gtr-t5-base,HotpotQA,ndcg_at_10,0.54926
|
39 |
+
gtr-t5-base,MSMARCO,ndcg_at_10,0.68745
|
40 |
+
gtr-t5-base,NFCorpus,ndcg_at_10,0.30223
|
41 |
+
gtr-t5-base,NQ,ndcg_at_10,0.50469
|
42 |
+
gtr-t5-base,QuoraRetrieval,ndcg_at_10,0.87977
|
43 |
+
gtr-t5-base,SCIDOCS,ndcg_at_10,0.14
|
44 |
+
gtr-t5-base,SciFact,ndcg_at_10,0.59738
|
45 |
+
gtr-t5-base,Touche2020,ndcg_at_10,0.25891
|
46 |
+
gtr-t5-base,TRECCOVID,ndcg_at_10,0.56047
|
47 |
+
gtr-t5-base,BIOSSES,cosine_spearman,0.7900420980306923
|
48 |
+
gtr-t5-base,SICK-R,cosine_spearman,0.7900420980306923
|
49 |
+
gtr-t5-base,STS12,cosine_spearman,0.6859358272758398
|
50 |
+
gtr-t5-base,STS13,cosine_spearman,0.7909019447848391
|
51 |
+
gtr-t5-base,STS14,cosine_spearman,0.7464141703710193
|
52 |
+
gtr-t5-base,STS15,cosine_spearman,0.8484757279716068
|
53 |
+
gtr-t5-base,STS16,cosine_spearman,0.8156700264541153
|
54 |
+
gtr-t5-base,STS17,cosine_spearman,0.8579615701004483
|
55 |
+
gtr-t5-base,STS22,cosine_spearman,0.6616950180914625
|
56 |
+
gtr-t5-base,STSBenchmark,cosine_spearman,0.7957572102026641
|
57 |
+
gtr-t5-base,SummEval,cosine_spearman,0.29669058153671907
|
gtr-t5-xxl_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6729850746268656
|
3 |
-
AmazonPolarityClassification,accuracy,0.7505425
|
4 |
-
AmazonReviewsClassification,accuracy,0.37298
|
5 |
-
Banking77Classification,accuracy,0.8232142857142858
|
6 |
-
EmotionClassification,accuracy,0.43190000000000006
|
7 |
-
ImdbClassification,accuracy,0.708028
|
8 |
-
MassiveIntentClassification,accuracy,0.7060860793544049
|
9 |
-
MassiveScenarioClassification,accuracy,0.7777404169468729
|
10 |
-
MTOPDomainClassification,accuracy,0.9384404924760602
|
11 |
-
MTOPIntentClassification,accuracy,0.6771317829457364
|
12 |
-
ToxicConversationsClassification,accuracy,0.684844
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5453593661573288
|
14 |
-
ArxivClusteringP2P,v_measure,0.37901257967755886
|
15 |
-
ArxivClusteringS2S,v_measure,0.32386812476732035
|
16 |
-
BiorxivClusteringP2P,v_measure,0.30479899128697724
|
17 |
-
BiorxivClusteringS2S,v_measure,0.274998862465458
|
18 |
-
MedrxivClusteringP2P,v_measure,0.29122279823498076
|
19 |
-
MedrxivClusteringS2S,v_measure,0.27559097352717216
|
20 |
-
RedditClustering,v_measure,0.6412600439035501
|
21 |
-
RedditClusteringP2P,v_measure,0.6284408311967189
|
22 |
-
StackExchangeClustering,v_measure,0.7143121835042072
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.32846363140242907
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.5043598055321865
|
25 |
-
SprintDuplicateQuestions,ap,0.9568031991336098
|
26 |
-
TwitterSemEval2015,ap,0.7754306712482096
|
27 |
-
TwitterURLCorpus,ap,0.851341871186482
|
28 |
-
AskUbuntuDupQuestions,map,0.632297938852618
|
29 |
-
MindSmallReranking,map,0.31927428563936344
|
30 |
-
SciDocsRR,map,0.7796354659897269
|
31 |
-
StackOverflowDupQuestions,map,0.5349865450479917
|
32 |
-
ArguAna,ndcg_at_10,0.53769
|
33 |
-
ClimateFEVER,ndcg_at_10,
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.3855908333333333
|
35 |
-
DBPedia,ndcg_at_10,
|
36 |
-
FEVER,ndcg_at_10,
|
37 |
-
FiQA2018,ndcg_at_10,0.46776
|
38 |
-
HotpotQA,ndcg_at_10,
|
39 |
-
MSMARCO,ndcg_at_10,
|
40 |
-
NFCorpus,ndcg_at_10,0.34179
|
41 |
-
NQ,ndcg_at_10,
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.89094
|
43 |
-
SCIDOCS,ndcg_at_10,0.15881
|
44 |
-
SciFact,ndcg_at_10,0.66772
|
45 |
-
Touche2020,ndcg_at_10,0.2676
|
46 |
-
TRECCOVID,ndcg_at_10,0.51903
|
47 |
-
BIOSSES,cosine_spearman,0.819072879998681
|
48 |
-
SICK-R,cosine_spearman,0.7429383974498337
|
49 |
-
STS12,cosine_spearman,0.701248199396564
|
50 |
-
STS13,cosine_spearman,0.8271549276218907
|
51 |
-
STS14,cosine_spearman,0.7824131910280874
|
52 |
-
STS15,cosine_spearman,0.8625849084541269
|
53 |
-
STS16,cosine_spearman,0.8161499741247533
|
54 |
-
STS17,cosine_spearman,0.8518454365763111
|
55 |
-
STS22,cosine_spearman,0.6576131885658799
|
56 |
-
STSBenchmark,cosine_spearman,0.7772574493807944
|
57 |
-
SummEval,cosine_spearman,0.3063683987345714
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
gtr-t5-xxl,AmazonCounterfactualClassification,accuracy,0.6729850746268656
|
3 |
+
gtr-t5-xxl,AmazonPolarityClassification,accuracy,0.7505425
|
4 |
+
gtr-t5-xxl,AmazonReviewsClassification,accuracy,0.37298
|
5 |
+
gtr-t5-xxl,Banking77Classification,accuracy,0.8232142857142858
|
6 |
+
gtr-t5-xxl,EmotionClassification,accuracy,0.43190000000000006
|
7 |
+
gtr-t5-xxl,ImdbClassification,accuracy,0.708028
|
8 |
+
gtr-t5-xxl,MassiveIntentClassification,accuracy,0.7060860793544049
|
9 |
+
gtr-t5-xxl,MassiveScenarioClassification,accuracy,0.7777404169468729
|
10 |
+
gtr-t5-xxl,MTOPDomainClassification,accuracy,0.9384404924760602
|
11 |
+
gtr-t5-xxl,MTOPIntentClassification,accuracy,0.6771317829457364
|
12 |
+
gtr-t5-xxl,ToxicConversationsClassification,accuracy,0.684844
|
13 |
+
gtr-t5-xxl,TweetSentimentExtractionClassification,accuracy,0.5453593661573288
|
14 |
+
gtr-t5-xxl,ArxivClusteringP2P,v_measure,0.37901257967755886
|
15 |
+
gtr-t5-xxl,ArxivClusteringS2S,v_measure,0.32386812476732035
|
16 |
+
gtr-t5-xxl,BiorxivClusteringP2P,v_measure,0.30479899128697724
|
17 |
+
gtr-t5-xxl,BiorxivClusteringS2S,v_measure,0.274998862465458
|
18 |
+
gtr-t5-xxl,MedrxivClusteringP2P,v_measure,0.29122279823498076
|
19 |
+
gtr-t5-xxl,MedrxivClusteringS2S,v_measure,0.27559097352717216
|
20 |
+
gtr-t5-xxl,RedditClustering,v_measure,0.6412600439035501
|
21 |
+
gtr-t5-xxl,RedditClusteringP2P,v_measure,0.6284408311967189
|
22 |
+
gtr-t5-xxl,StackExchangeClustering,v_measure,0.7143121835042072
|
23 |
+
gtr-t5-xxl,StackExchangeClusteringP2P,v_measure,0.32846363140242907
|
24 |
+
gtr-t5-xxl,TwentyNewsgroupsClustering,v_measure,0.5043598055321865
|
25 |
+
gtr-t5-xxl,SprintDuplicateQuestions,ap,0.9568031991336098
|
26 |
+
gtr-t5-xxl,TwitterSemEval2015,ap,0.7754306712482096
|
27 |
+
gtr-t5-xxl,TwitterURLCorpus,ap,0.851341871186482
|
28 |
+
gtr-t5-xxl,AskUbuntuDupQuestions,map,0.632297938852618
|
29 |
+
gtr-t5-xxl,MindSmallReranking,map,0.31927428563936344
|
30 |
+
gtr-t5-xxl,SciDocsRR,map,0.7796354659897269
|
31 |
+
gtr-t5-xxl,StackOverflowDupQuestions,map,0.5349865450479917
|
32 |
+
gtr-t5-xxl,ArguAna,ndcg_at_10,0.53769
|
33 |
+
gtr-t5-xxl,ClimateFEVER,ndcg_at_10,
|
34 |
+
gtr-t5-xxl,CQADupstackRetrieval,ndcg_at_10,0.3855908333333333
|
35 |
+
gtr-t5-xxl,DBPedia,ndcg_at_10,
|
36 |
+
gtr-t5-xxl,FEVER,ndcg_at_10,
|
37 |
+
gtr-t5-xxl,FiQA2018,ndcg_at_10,0.46776
|
38 |
+
gtr-t5-xxl,HotpotQA,ndcg_at_10,
|
39 |
+
gtr-t5-xxl,MSMARCO,ndcg_at_10,
|
40 |
+
gtr-t5-xxl,NFCorpus,ndcg_at_10,0.34179
|
41 |
+
gtr-t5-xxl,NQ,ndcg_at_10,
|
42 |
+
gtr-t5-xxl,QuoraRetrieval,ndcg_at_10,0.89094
|
43 |
+
gtr-t5-xxl,SCIDOCS,ndcg_at_10,0.15881
|
44 |
+
gtr-t5-xxl,SciFact,ndcg_at_10,0.66772
|
45 |
+
gtr-t5-xxl,Touche2020,ndcg_at_10,0.2676
|
46 |
+
gtr-t5-xxl,TRECCOVID,ndcg_at_10,0.51903
|
47 |
+
gtr-t5-xxl,BIOSSES,cosine_spearman,0.819072879998681
|
48 |
+
gtr-t5-xxl,SICK-R,cosine_spearman,0.7429383974498337
|
49 |
+
gtr-t5-xxl,STS12,cosine_spearman,0.701248199396564
|
50 |
+
gtr-t5-xxl,STS13,cosine_spearman,0.8271549276218907
|
51 |
+
gtr-t5-xxl,STS14,cosine_spearman,0.7824131910280874
|
52 |
+
gtr-t5-xxl,STS15,cosine_spearman,0.8625849084541269
|
53 |
+
gtr-t5-xxl,STS16,cosine_spearman,0.8161499741247533
|
54 |
+
gtr-t5-xxl,STS17,cosine_spearman,0.8518454365763111
|
55 |
+
gtr-t5-xxl,STS22,cosine_spearman,0.6576131885658799
|
56 |
+
gtr-t5-xxl,STSBenchmark,cosine_spearman,0.7772574493807944
|
57 |
+
gtr-t5-xxl,SummEval,cosine_spearman,0.3063683987345714
|
komninos_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6053731343283582
|
3 |
-
AmazonPolarityClassification,accuracy,0.5958574999999999
|
4 |
-
AmazonReviewsClassification,accuracy,0.31013999999999997
|
5 |
-
Banking77Classification,accuracy,0.6704870129870131
|
6 |
-
EmotionClassification,accuracy,0.3318
|
7 |
-
ImdbClassification,accuracy,0.639756
|
8 |
-
MassiveIntentClassification,accuracy,0.5721250840618696
|
9 |
-
MassiveScenarioClassification,accuracy,0.6611297915265636
|
10 |
-
MTOPDomainClassification,accuracy,0.7857045143638851
|
11 |
-
MTOPIntentClassification,accuracy,0.5707250341997263
|
12 |
-
ToxicConversationsClassification,accuracy,0.677564
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.4968307866440294
|
14 |
-
ArxivClusteringP2P,v_measure,0.3472624365591666
|
15 |
-
ArxivClusteringS2S,v_measure,0.26008268382733774
|
16 |
-
BiorxivClusteringP2P,v_measure,0.2975913127921992
|
17 |
-
BiorxivClusteringS2S,v_measure,0.20711732977389513
|
18 |
-
MedrxivClusteringP2P,v_measure,0.2665246941866784
|
19 |
-
MedrxivClusteringS2S,v_measure,0.21503100153759527
|
20 |
-
RedditClustering,v_measure,0.288407473136122
|
21 |
-
RedditClusteringP2P,v_measure,0.0737029180225715
|
22 |
-
StackExchangeClustering,v_measure,0.39038573398030996
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3022655763114615
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.27420294690851976
|
25 |
-
SprintDuplicateQuestions,ap,0.8555090397855645
|
26 |
-
TwitterSemEval2015,ap,0.5385004776679795
|
27 |
-
TwitterURLCorpus,ap,0.7941032575876537
|
28 |
-
AskUbuntuDupQuestions,map,0.5087892970617054
|
29 |
-
MindSmallReranking,map,0.28921918247348344
|
30 |
-
SciDocsRR,map,0.6355311507513624
|
31 |
-
StackOverflowDupQuestions,map,0.3564753567606952
|
32 |
-
ArguAna,ndcg_at_10,0.30959
|
33 |
-
ClimateFEVER,ndcg_at_10,0.14867
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.16788249999999996
|
35 |
-
DBPedia,ndcg_at_10,0.15877
|
36 |
-
FEVER,ndcg_at_10,0.15558
|
37 |
-
FiQA2018,ndcg_at_10,0.10488
|
38 |
-
HotpotQA,ndcg_at_10,0.20768
|
39 |
-
MSMARCO,ndcg_at_10,0.28197
|
40 |
-
NFCorpus,ndcg_at_10,0.11787
|
41 |
-
NQ,ndcg_at_10,0.12751
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.71575
|
43 |
-
SCIDOCS,ndcg_at_10,0.0847
|
44 |
-
SciFact,ndcg_at_10,0.29526
|
45 |
-
Touche2020,ndcg_at_10,0.13173
|
46 |
-
TRECCOVID,ndcg_at_10,0.35923
|
47 |
-
BIOSSES,cosine_spearman,0.502481328896023
|
48 |
-
SICK-R,cosine_spearman,0.5548989810259632
|
49 |
-
STS12,cosine_spearman,0.5351287401411937
|
50 |
-
STS13,cosine_spearman,0.7080105270649192
|
51 |
-
STS14,cosine_spearman,0.6356156596006354
|
52 |
-
STS15,cosine_spearman,0.740812262403237
|
53 |
-
STS16,cosine_spearman,0.646011290069885
|
54 |
-
STS17,cosine_spearman,0.7691005426258415
|
55 |
-
STS22,cosine_spearman,0.5389239838658062
|
56 |
-
STSBenchmark,cosine_spearman,0.6155281748776228
|
57 |
-
SummEval,cosine_spearman,0.3049084135949722
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
komninos,AmazonCounterfactualClassification,accuracy,0.6053731343283582
|
3 |
+
komninos,AmazonPolarityClassification,accuracy,0.5958574999999999
|
4 |
+
komninos,AmazonReviewsClassification,accuracy,0.31013999999999997
|
5 |
+
komninos,Banking77Classification,accuracy,0.6704870129870131
|
6 |
+
komninos,EmotionClassification,accuracy,0.3318
|
7 |
+
komninos,ImdbClassification,accuracy,0.639756
|
8 |
+
komninos,MassiveIntentClassification,accuracy,0.5721250840618696
|
9 |
+
komninos,MassiveScenarioClassification,accuracy,0.6611297915265636
|
10 |
+
komninos,MTOPDomainClassification,accuracy,0.7857045143638851
|
11 |
+
komninos,MTOPIntentClassification,accuracy,0.5707250341997263
|
12 |
+
komninos,ToxicConversationsClassification,accuracy,0.677564
|
13 |
+
komninos,TweetSentimentExtractionClassification,accuracy,0.4968307866440294
|
14 |
+
komninos,ArxivClusteringP2P,v_measure,0.3472624365591666
|
15 |
+
komninos,ArxivClusteringS2S,v_measure,0.26008268382733774
|
16 |
+
komninos,BiorxivClusteringP2P,v_measure,0.2975913127921992
|
17 |
+
komninos,BiorxivClusteringS2S,v_measure,0.20711732977389513
|
18 |
+
komninos,MedrxivClusteringP2P,v_measure,0.2665246941866784
|
19 |
+
komninos,MedrxivClusteringS2S,v_measure,0.21503100153759527
|
20 |
+
komninos,RedditClustering,v_measure,0.288407473136122
|
21 |
+
komninos,RedditClusteringP2P,v_measure,0.0737029180225715
|
22 |
+
komninos,StackExchangeClustering,v_measure,0.39038573398030996
|
23 |
+
komninos,StackExchangeClusteringP2P,v_measure,0.3022655763114615
|
24 |
+
komninos,TwentyNewsgroupsClustering,v_measure,0.27420294690851976
|
25 |
+
komninos,SprintDuplicateQuestions,ap,0.8555090397855645
|
26 |
+
komninos,TwitterSemEval2015,ap,0.5385004776679795
|
27 |
+
komninos,TwitterURLCorpus,ap,0.7941032575876537
|
28 |
+
komninos,AskUbuntuDupQuestions,map,0.5087892970617054
|
29 |
+
komninos,MindSmallReranking,map,0.28921918247348344
|
30 |
+
komninos,SciDocsRR,map,0.6355311507513624
|
31 |
+
komninos,StackOverflowDupQuestions,map,0.3564753567606952
|
32 |
+
komninos,ArguAna,ndcg_at_10,0.30959
|
33 |
+
komninos,ClimateFEVER,ndcg_at_10,0.14867
|
34 |
+
komninos,CQADupstackRetrieval,ndcg_at_10,0.16788249999999996
|
35 |
+
komninos,DBPedia,ndcg_at_10,0.15877
|
36 |
+
komninos,FEVER,ndcg_at_10,0.15558
|
37 |
+
komninos,FiQA2018,ndcg_at_10,0.10488
|
38 |
+
komninos,HotpotQA,ndcg_at_10,0.20768
|
39 |
+
komninos,MSMARCO,ndcg_at_10,0.28197
|
40 |
+
komninos,NFCorpus,ndcg_at_10,0.11787
|
41 |
+
komninos,NQ,ndcg_at_10,0.12751
|
42 |
+
komninos,QuoraRetrieval,ndcg_at_10,0.71575
|
43 |
+
komninos,SCIDOCS,ndcg_at_10,0.0847
|
44 |
+
komninos,SciFact,ndcg_at_10,0.29526
|
45 |
+
komninos,Touche2020,ndcg_at_10,0.13173
|
46 |
+
komninos,TRECCOVID,ndcg_at_10,0.35923
|
47 |
+
komninos,BIOSSES,cosine_spearman,0.502481328896023
|
48 |
+
komninos,SICK-R,cosine_spearman,0.5548989810259632
|
49 |
+
komninos,STS12,cosine_spearman,0.5351287401411937
|
50 |
+
komninos,STS13,cosine_spearman,0.7080105270649192
|
51 |
+
komninos,STS14,cosine_spearman,0.6356156596006354
|
52 |
+
komninos,STS15,cosine_spearman,0.740812262403237
|
53 |
+
komninos,STS16,cosine_spearman,0.646011290069885
|
54 |
+
komninos,STS17,cosine_spearman,0.7691005426258415
|
55 |
+
komninos,STS22,cosine_spearman,0.5389239838658062
|
56 |
+
komninos,STSBenchmark,cosine_spearman,0.6155281748776228
|
57 |
+
komninos,SummEval,cosine_spearman,0.3049084135949722
|
msmarco-bert-co-condensor_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6405970149253731
|
3 |
-
AmazonPolarityClassification,accuracy,0.66883225
|
4 |
-
AmazonReviewsClassification,accuracy,0.34852
|
5 |
-
Banking77Classification,accuracy,0.8235389610389611
|
6 |
-
EmotionClassification,accuracy,0.4191
|
7 |
-
ImdbClassification,accuracy,0.6016519999999999
|
8 |
-
MassiveIntentClassification,accuracy,0.7040013449899125
|
9 |
-
MassiveScenarioClassification,accuracy,0.73728984532616
|
10 |
-
MTOPDomainClassification,accuracy,0.9133834929320566
|
11 |
-
MTOPIntentClassification,accuracy,0.7106931144550843
|
12 |
-
ToxicConversationsClassification,accuracy,0.6400720000000001
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5573571024335031
|
14 |
-
ArxivClusteringP2P,v_measure,0.369429585124196
|
15 |
-
ArxivClusteringS2S,v_measure,0.2903421184003919
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3234945427009579
|
17 |
-
BiorxivClusteringS2S,v_measure,0.28155732412289625
|
18 |
-
MedrxivClusteringP2P,v_measure,0.3023226235842047
|
19 |
-
MedrxivClusteringS2S,v_measure,0.2700892040743902
|
20 |
-
RedditClustering,v_measure,0.48044277851861944
|
21 |
-
RedditClusteringP2P,v_measure,0.5352963896499805
|
22 |
-
StackExchangeClustering,v_measure,0.5953788370911354
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.30481638824199886
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.3867893425770875
|
25 |
-
SprintDuplicateQuestions,ap,0.9609255002552
|
26 |
-
TwitterSemEval2015,ap,0.6594999377083578
|
27 |
-
TwitterURLCorpus,ap,0.8317318134999252
|
28 |
-
AskUbuntuDupQuestions,map,0.5898600276337803
|
29 |
-
MindSmallReranking,map,0.2713298598749459
|
30 |
-
SciDocsRR,map,0.7277578387510362
|
31 |
-
StackOverflowDupQuestions,map,0.4848067606975988
|
32 |
-
ArguAna,ndcg_at_10,0.45154
|
33 |
-
ClimateFEVER,ndcg_at_10,0.16957
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.2771733333333333
|
35 |
-
DBPedia,ndcg_at_10,0.27859
|
36 |
-
FEVER,ndcg_at_10,0.45684
|
37 |
-
FiQA2018,ndcg_at_10,0.15619
|
38 |
-
HotpotQA,ndcg_at_10,0.35608
|
39 |
-
MSMARCO,ndcg_at_10,0.56235
|
40 |
-
NFCorpus,ndcg_at_10,0.22291
|
41 |
-
NQ,ndcg_at_10,0.29855
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.86506
|
43 |
-
SCIDOCS,ndcg_at_10,0.10131
|
44 |
-
SciFact,ndcg_at_10,0.52311
|
45 |
-
Touche2020,ndcg_at_10,0.08573
|
46 |
-
TRECCOVID,ndcg_at_10,0.40536
|
47 |
-
BIOSSES,cosine_spearman,0.7731501468710659
|
48 |
-
SICK-R,cosine_spearman,0.719994667180189
|
49 |
-
STS12,cosine_spearman,0.6819498646316081
|
50 |
-
STS13,cosine_spearman,0.8040441089650371
|
51 |
-
STS14,cosine_spearman,0.7401841404523465
|
52 |
-
STS15,cosine_spearman,0.8257380490897778
|
53 |
-
STS16,cosine_spearman,0.7977999310698289
|
54 |
-
STS17,cosine_spearman,0.8593569295808465
|
55 |
-
STS22,cosine_spearman,0.675433916658919
|
56 |
-
STSBenchmark,cosine_spearman,0.7696651521767748
|
57 |
-
SummEval,cosine_spearman,0.29504172709559223
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
msmarco-bert-co-condensor,AmazonCounterfactualClassification,accuracy,0.6405970149253731
|
3 |
+
msmarco-bert-co-condensor,AmazonPolarityClassification,accuracy,0.66883225
|
4 |
+
msmarco-bert-co-condensor,AmazonReviewsClassification,accuracy,0.34852
|
5 |
+
msmarco-bert-co-condensor,Banking77Classification,accuracy,0.8235389610389611
|
6 |
+
msmarco-bert-co-condensor,EmotionClassification,accuracy,0.4191
|
7 |
+
msmarco-bert-co-condensor,ImdbClassification,accuracy,0.6016519999999999
|
8 |
+
msmarco-bert-co-condensor,MassiveIntentClassification,accuracy,0.7040013449899125
|
9 |
+
msmarco-bert-co-condensor,MassiveScenarioClassification,accuracy,0.73728984532616
|
10 |
+
msmarco-bert-co-condensor,MTOPDomainClassification,accuracy,0.9133834929320566
|
11 |
+
msmarco-bert-co-condensor,MTOPIntentClassification,accuracy,0.7106931144550843
|
12 |
+
msmarco-bert-co-condensor,ToxicConversationsClassification,accuracy,0.6400720000000001
|
13 |
+
msmarco-bert-co-condensor,TweetSentimentExtractionClassification,accuracy,0.5573571024335031
|
14 |
+
msmarco-bert-co-condensor,ArxivClusteringP2P,v_measure,0.369429585124196
|
15 |
+
msmarco-bert-co-condensor,ArxivClusteringS2S,v_measure,0.2903421184003919
|
16 |
+
msmarco-bert-co-condensor,BiorxivClusteringP2P,v_measure,0.3234945427009579
|
17 |
+
msmarco-bert-co-condensor,BiorxivClusteringS2S,v_measure,0.28155732412289625
|
18 |
+
msmarco-bert-co-condensor,MedrxivClusteringP2P,v_measure,0.3023226235842047
|
19 |
+
msmarco-bert-co-condensor,MedrxivClusteringS2S,v_measure,0.2700892040743902
|
20 |
+
msmarco-bert-co-condensor,RedditClustering,v_measure,0.48044277851861944
|
21 |
+
msmarco-bert-co-condensor,RedditClusteringP2P,v_measure,0.5352963896499805
|
22 |
+
msmarco-bert-co-condensor,StackExchangeClustering,v_measure,0.5953788370911354
|
23 |
+
msmarco-bert-co-condensor,StackExchangeClusteringP2P,v_measure,0.30481638824199886
|
24 |
+
msmarco-bert-co-condensor,TwentyNewsgroupsClustering,v_measure,0.3867893425770875
|
25 |
+
msmarco-bert-co-condensor,SprintDuplicateQuestions,ap,0.9609255002552
|
26 |
+
msmarco-bert-co-condensor,TwitterSemEval2015,ap,0.6594999377083578
|
27 |
+
msmarco-bert-co-condensor,TwitterURLCorpus,ap,0.8317318134999252
|
28 |
+
msmarco-bert-co-condensor,AskUbuntuDupQuestions,map,0.5898600276337803
|
29 |
+
msmarco-bert-co-condensor,MindSmallReranking,map,0.2713298598749459
|
30 |
+
msmarco-bert-co-condensor,SciDocsRR,map,0.7277578387510362
|
31 |
+
msmarco-bert-co-condensor,StackOverflowDupQuestions,map,0.4848067606975988
|
32 |
+
msmarco-bert-co-condensor,ArguAna,ndcg_at_10,0.45154
|
33 |
+
msmarco-bert-co-condensor,ClimateFEVER,ndcg_at_10,0.16957
|
34 |
+
msmarco-bert-co-condensor,CQADupstackRetrieval,ndcg_at_10,0.2771733333333333
|
35 |
+
msmarco-bert-co-condensor,DBPedia,ndcg_at_10,0.27859
|
36 |
+
msmarco-bert-co-condensor,FEVER,ndcg_at_10,0.45684
|
37 |
+
msmarco-bert-co-condensor,FiQA2018,ndcg_at_10,0.15619
|
38 |
+
msmarco-bert-co-condensor,HotpotQA,ndcg_at_10,0.35608
|
39 |
+
msmarco-bert-co-condensor,MSMARCO,ndcg_at_10,0.56235
|
40 |
+
msmarco-bert-co-condensor,NFCorpus,ndcg_at_10,0.22291
|
41 |
+
msmarco-bert-co-condensor,NQ,ndcg_at_10,0.29855
|
42 |
+
msmarco-bert-co-condensor,QuoraRetrieval,ndcg_at_10,0.86506
|
43 |
+
msmarco-bert-co-condensor,SCIDOCS,ndcg_at_10,0.10131
|
44 |
+
msmarco-bert-co-condensor,SciFact,ndcg_at_10,0.52311
|
45 |
+
msmarco-bert-co-condensor,Touche2020,ndcg_at_10,0.08573
|
46 |
+
msmarco-bert-co-condensor,TRECCOVID,ndcg_at_10,0.40536
|
47 |
+
msmarco-bert-co-condensor,BIOSSES,cosine_spearman,0.7731501468710659
|
48 |
+
msmarco-bert-co-condensor,SICK-R,cosine_spearman,0.719994667180189
|
49 |
+
msmarco-bert-co-condensor,STS12,cosine_spearman,0.6819498646316081
|
50 |
+
msmarco-bert-co-condensor,STS13,cosine_spearman,0.8040441089650371
|
51 |
+
msmarco-bert-co-condensor,STS14,cosine_spearman,0.7401841404523465
|
52 |
+
msmarco-bert-co-condensor,STS15,cosine_spearman,0.8257380490897778
|
53 |
+
msmarco-bert-co-condensor,STS16,cosine_spearman,0.7977999310698289
|
54 |
+
msmarco-bert-co-condensor,STS17,cosine_spearman,0.8593569295808465
|
55 |
+
msmarco-bert-co-condensor,STS22,cosine_spearman,0.675433916658919
|
56 |
+
msmarco-bert-co-condensor,STSBenchmark,cosine_spearman,0.7696651521767748
|
57 |
+
msmarco-bert-co-condensor,SummEval,cosine_spearman,0.29504172709559223
|
sentence-t5-base_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7582089552238807
|
3 |
-
AmazonPolarityClassification,accuracy,0.8511737500000001
|
4 |
-
AmazonReviewsClassification,accuracy,0.44943999999999995
|
5 |
-
Banking77Classification,accuracy,0.7647727272727273
|
6 |
-
EmotionClassification,accuracy,0.5135500000000001
|
7 |
-
ImdbClassification,accuracy,0.773372
|
8 |
-
MassiveIntentClassification,accuracy,0.6974445191661063
|
9 |
-
MassiveScenarioClassification,accuracy,0.7232347007397445
|
10 |
-
MTOPDomainClassification,accuracy,0.903374373005016
|
11 |
-
MTOPIntentClassification,accuracy,0.6332421340629275
|
12 |
-
ToxicConversationsClassification,accuracy,0.6819919999999999
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.6271080928126768
|
14 |
-
ArxivClusteringP2P,v_measure,0.39275291662236395
|
15 |
-
ArxivClusteringS2S,v_measure,0.27261128959373326
|
16 |
-
BiorxivClusteringP2P,v_measure,0.33985330956534987
|
17 |
-
BiorxivClusteringS2S,v_measure,0.22921149832439514
|
18 |
-
MedrxivClusteringP2P,v_measure,0.33201641285078753
|
19 |
-
MedrxivClusteringS2S,v_measure,0.2613067187536383
|
20 |
-
RedditClustering,v_measure,0.5293053483970847
|
21 |
-
RedditClusteringP2P,v_measure,0.596686566444821
|
22 |
-
StackExchangeClustering,v_measure,0.6313072579524015
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3568114137558022
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.48099101871064487
|
25 |
-
SprintDuplicateQuestions,ap,0.9122814257221482
|
26 |
-
TwitterSemEval2015,ap,0.7825096573546108
|
27 |
-
TwitterURLCorpus,ap,0.8604575990028458
|
28 |
-
AskUbuntuDupQuestions,map,0.5972875661091822
|
29 |
-
MindSmallReranking,map,0.30196798710053224
|
30 |
-
SciDocsRR,map,0.7396228332723271
|
31 |
-
StackOverflowDupQuestions,map,0.4846269194141537
|
32 |
-
ArguAna,ndcg_at_10,0.44846
|
33 |
-
ClimateFEVER,ndcg_at_10,0.10367
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.3523175
|
35 |
-
DBPedia,ndcg_at_10,0.27766
|
36 |
-
FEVER,ndcg_at_10,0.26165
|
37 |
-
FiQA2018,ndcg_at_10,0.34832
|
38 |
-
HotpotQA,ndcg_at_10,0.33198
|
39 |
-
MSMARCO,ndcg_at_10,0.4111
|
40 |
-
NFCorpus,ndcg_at_10,0.28645
|
41 |
-
NQ,ndcg_at_10,0.36324
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.85491
|
43 |
-
SCIDOCS,ndcg_at_10,0.14155
|
44 |
-
SciFact,ndcg_at_10,0.4576
|
45 |
-
Touche2020,ndcg_at_10,0.20296
|
46 |
-
TRECCOVID,ndcg_at_10,0.40699
|
47 |
-
BIOSSES,cosine_spearman,0.7589088585182279
|
48 |
-
SICK-R,cosine_spearman,0.7589088585182279
|
49 |
-
STS12,cosine_spearman,0.780511871449349
|
50 |
-
STS13,cosine_spearman,0.8584897342040492
|
51 |
-
STS14,cosine_spearman,0.8218926664662587
|
52 |
-
STS15,cosine_spearman,0.8745860981918768
|
53 |
-
STS16,cosine_spearman,0.8403233567112526
|
54 |
-
STS17,cosine_spearman,0.8956885540021488
|
55 |
-
STS22,cosine_spearman,0.6265994888539158
|
56 |
-
STSBenchmark,cosine_spearman,0.8552030817522575
|
57 |
-
SummEval,cosine_spearman,0.313940211538976
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
sentence-t5-base,AmazonCounterfactualClassification,accuracy,0.7582089552238807
|
3 |
+
sentence-t5-base,AmazonPolarityClassification,accuracy,0.8511737500000001
|
4 |
+
sentence-t5-base,AmazonReviewsClassification,accuracy,0.44943999999999995
|
5 |
+
sentence-t5-base,Banking77Classification,accuracy,0.7647727272727273
|
6 |
+
sentence-t5-base,EmotionClassification,accuracy,0.5135500000000001
|
7 |
+
sentence-t5-base,ImdbClassification,accuracy,0.773372
|
8 |
+
sentence-t5-base,MassiveIntentClassification,accuracy,0.6974445191661063
|
9 |
+
sentence-t5-base,MassiveScenarioClassification,accuracy,0.7232347007397445
|
10 |
+
sentence-t5-base,MTOPDomainClassification,accuracy,0.903374373005016
|
11 |
+
sentence-t5-base,MTOPIntentClassification,accuracy,0.6332421340629275
|
12 |
+
sentence-t5-base,ToxicConversationsClassification,accuracy,0.6819919999999999
|
13 |
+
sentence-t5-base,TweetSentimentExtractionClassification,accuracy,0.6271080928126768
|
14 |
+
sentence-t5-base,ArxivClusteringP2P,v_measure,0.39275291662236395
|
15 |
+
sentence-t5-base,ArxivClusteringS2S,v_measure,0.27261128959373326
|
16 |
+
sentence-t5-base,BiorxivClusteringP2P,v_measure,0.33985330956534987
|
17 |
+
sentence-t5-base,BiorxivClusteringS2S,v_measure,0.22921149832439514
|
18 |
+
sentence-t5-base,MedrxivClusteringP2P,v_measure,0.33201641285078753
|
19 |
+
sentence-t5-base,MedrxivClusteringS2S,v_measure,0.2613067187536383
|
20 |
+
sentence-t5-base,RedditClustering,v_measure,0.5293053483970847
|
21 |
+
sentence-t5-base,RedditClusteringP2P,v_measure,0.596686566444821
|
22 |
+
sentence-t5-base,StackExchangeClustering,v_measure,0.6313072579524015
|
23 |
+
sentence-t5-base,StackExchangeClusteringP2P,v_measure,0.3568114137558022
|
24 |
+
sentence-t5-base,TwentyNewsgroupsClustering,v_measure,0.48099101871064487
|
25 |
+
sentence-t5-base,SprintDuplicateQuestions,ap,0.9122814257221482
|
26 |
+
sentence-t5-base,TwitterSemEval2015,ap,0.7825096573546108
|
27 |
+
sentence-t5-base,TwitterURLCorpus,ap,0.8604575990028458
|
28 |
+
sentence-t5-base,AskUbuntuDupQuestions,map,0.5972875661091822
|
29 |
+
sentence-t5-base,MindSmallReranking,map,0.30196798710053224
|
30 |
+
sentence-t5-base,SciDocsRR,map,0.7396228332723271
|
31 |
+
sentence-t5-base,StackOverflowDupQuestions,map,0.4846269194141537
|
32 |
+
sentence-t5-base,ArguAna,ndcg_at_10,0.44846
|
33 |
+
sentence-t5-base,ClimateFEVER,ndcg_at_10,0.10367
|
34 |
+
sentence-t5-base,CQADupstackRetrieval,ndcg_at_10,0.3523175
|
35 |
+
sentence-t5-base,DBPedia,ndcg_at_10,0.27766
|
36 |
+
sentence-t5-base,FEVER,ndcg_at_10,0.26165
|
37 |
+
sentence-t5-base,FiQA2018,ndcg_at_10,0.34832
|
38 |
+
sentence-t5-base,HotpotQA,ndcg_at_10,0.33198
|
39 |
+
sentence-t5-base,MSMARCO,ndcg_at_10,0.4111
|
40 |
+
sentence-t5-base,NFCorpus,ndcg_at_10,0.28645
|
41 |
+
sentence-t5-base,NQ,ndcg_at_10,0.36324
|
42 |
+
sentence-t5-base,QuoraRetrieval,ndcg_at_10,0.85491
|
43 |
+
sentence-t5-base,SCIDOCS,ndcg_at_10,0.14155
|
44 |
+
sentence-t5-base,SciFact,ndcg_at_10,0.4576
|
45 |
+
sentence-t5-base,Touche2020,ndcg_at_10,0.20296
|
46 |
+
sentence-t5-base,TRECCOVID,ndcg_at_10,0.40699
|
47 |
+
sentence-t5-base,BIOSSES,cosine_spearman,0.7589088585182279
|
48 |
+
sentence-t5-base,SICK-R,cosine_spearman,0.7589088585182279
|
49 |
+
sentence-t5-base,STS12,cosine_spearman,0.780511871449349
|
50 |
+
sentence-t5-base,STS13,cosine_spearman,0.8584897342040492
|
51 |
+
sentence-t5-base,STS14,cosine_spearman,0.8218926664662587
|
52 |
+
sentence-t5-base,STS15,cosine_spearman,0.8745860981918768
|
53 |
+
sentence-t5-base,STS16,cosine_spearman,0.8403233567112526
|
54 |
+
sentence-t5-base,STS17,cosine_spearman,0.8956885540021488
|
55 |
+
sentence-t5-base,STS22,cosine_spearman,0.6265994888539158
|
56 |
+
sentence-t5-base,STSBenchmark,cosine_spearman,0.8552030817522575
|
57 |
+
sentence-t5-base,SummEval,cosine_spearman,0.313940211538976
|
sentence-t5-xxl_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7707462686567165
|
3 |
-
AmazonPolarityClassification,accuracy,0.9278587500000001
|
4 |
-
AmazonReviewsClassification,accuracy,0.48926
|
5 |
-
Banking77Classification,accuracy,0.8230844155844155
|
6 |
-
EmotionClassification,accuracy,0.4857
|
7 |
-
ImdbClassification,accuracy,0.902268
|
8 |
-
MassiveIntentClassification,accuracy,0.7344317417619368
|
9 |
-
MassiveScenarioClassification,accuracy,0.7481842636180229
|
10 |
-
MTOPDomainClassification,accuracy,0.9249430004559963
|
11 |
-
MTOPIntentClassification,accuracy,0.683264933880529
|
12 |
-
ToxicConversationsClassification,accuracy,0.700366
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.6200905489530277
|
14 |
-
ArxivClusteringP2P,v_measure,0.428912091513744
|
15 |
-
ArxivClusteringS2S,v_measure,0.334692173332607
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3652762269967934
|
17 |
-
BiorxivClusteringS2S,v_measure,0.2866305685584199
|
18 |
-
MedrxivClusteringP2P,v_measure,0.3208579812374771
|
19 |
-
MedrxivClusteringS2S,v_measure,0.2681601616447503
|
20 |
-
RedditClustering,v_measure,0.5898542282844221
|
21 |
-
RedditClusteringP2P,v_measure,0.6445543643659679
|
22 |
-
StackExchangeClustering,v_measure,0.7077797692376002
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.3525246972831145
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.5093024400468489
|
25 |
-
SprintDuplicateQuestions,ap,0.8888675641034331
|
26 |
-
TwitterSemEval2015,ap,0.802843387858549
|
27 |
-
TwitterURLCorpus,ap,0.8601198098811131
|
28 |
-
AskUbuntuDupQuestions,map,0.6615532012360636
|
29 |
-
MindSmallReranking,map,0.3059647884122016
|
30 |
-
SciDocsRR,map,0.7609412848204858
|
31 |
-
StackOverflowDupQuestions,map,0.5285484177221984
|
32 |
-
ArguAna,ndcg_at_10,0.39847
|
33 |
-
ClimateFEVER,ndcg_at_10,
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.44654833333333344
|
35 |
-
DBPedia,ndcg_at_10,
|
36 |
-
FEVER,ndcg_at_10,
|
37 |
-
FiQA2018,ndcg_at_10,0.46677
|
38 |
-
HotpotQA,ndcg_at_10,
|
39 |
-
MSMARCO,ndcg_at_10,
|
40 |
-
NFCorpus,ndcg_at_10,0.35077
|
41 |
-
NQ,ndcg_at_10,0.5287
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.85959
|
43 |
-
SCIDOCS,ndcg_at_10,0.17173
|
44 |
-
SciFact,ndcg_at_10,0.5538
|
45 |
-
Touche2020,ndcg_at_10,0.21647
|
46 |
-
TRECCOVID,ndcg_at_10,0.59481
|
47 |
-
BIOSSES,cosine_spearman,0.8042954348508337
|
48 |
-
SICK-R,cosine_spearman,0.8042954348508337
|
49 |
-
STS12,cosine_spearman,0.7884633095838921
|
50 |
-
STS13,cosine_spearman,0.8893834486158684
|
51 |
-
STS14,cosine_spearman,0.8485684801116044
|
52 |
-
STS15,cosine_spearman,0.8931982869751905
|
53 |
-
STS16,cosine_spearman,0.846657396318755
|
54 |
-
STS17,cosine_spearman,0.8946272351246337
|
55 |
-
STS22,cosine_spearman,0.6532938965959216
|
56 |
-
STSBenchmark,cosine_spearman,0.8401139141474803
|
57 |
-
SummEval,cosine_spearman,0.30077056036328337
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
sentence-t5-xxl,AmazonCounterfactualClassification,accuracy,0.7707462686567165
|
3 |
+
sentence-t5-xxl,AmazonPolarityClassification,accuracy,0.9278587500000001
|
4 |
+
sentence-t5-xxl,AmazonReviewsClassification,accuracy,0.48926
|
5 |
+
sentence-t5-xxl,Banking77Classification,accuracy,0.8230844155844155
|
6 |
+
sentence-t5-xxl,EmotionClassification,accuracy,0.4857
|
7 |
+
sentence-t5-xxl,ImdbClassification,accuracy,0.902268
|
8 |
+
sentence-t5-xxl,MassiveIntentClassification,accuracy,0.7344317417619368
|
9 |
+
sentence-t5-xxl,MassiveScenarioClassification,accuracy,0.7481842636180229
|
10 |
+
sentence-t5-xxl,MTOPDomainClassification,accuracy,0.9249430004559963
|
11 |
+
sentence-t5-xxl,MTOPIntentClassification,accuracy,0.683264933880529
|
12 |
+
sentence-t5-xxl,ToxicConversationsClassification,accuracy,0.700366
|
13 |
+
sentence-t5-xxl,TweetSentimentExtractionClassification,accuracy,0.6200905489530277
|
14 |
+
sentence-t5-xxl,ArxivClusteringP2P,v_measure,0.428912091513744
|
15 |
+
sentence-t5-xxl,ArxivClusteringS2S,v_measure,0.334692173332607
|
16 |
+
sentence-t5-xxl,BiorxivClusteringP2P,v_measure,0.3652762269967934
|
17 |
+
sentence-t5-xxl,BiorxivClusteringS2S,v_measure,0.2866305685584199
|
18 |
+
sentence-t5-xxl,MedrxivClusteringP2P,v_measure,0.3208579812374771
|
19 |
+
sentence-t5-xxl,MedrxivClusteringS2S,v_measure,0.2681601616447503
|
20 |
+
sentence-t5-xxl,RedditClustering,v_measure,0.5898542282844221
|
21 |
+
sentence-t5-xxl,RedditClusteringP2P,v_measure,0.6445543643659679
|
22 |
+
sentence-t5-xxl,StackExchangeClustering,v_measure,0.7077797692376002
|
23 |
+
sentence-t5-xxl,StackExchangeClusteringP2P,v_measure,0.3525246972831145
|
24 |
+
sentence-t5-xxl,TwentyNewsgroupsClustering,v_measure,0.5093024400468489
|
25 |
+
sentence-t5-xxl,SprintDuplicateQuestions,ap,0.8888675641034331
|
26 |
+
sentence-t5-xxl,TwitterSemEval2015,ap,0.802843387858549
|
27 |
+
sentence-t5-xxl,TwitterURLCorpus,ap,0.8601198098811131
|
28 |
+
sentence-t5-xxl,AskUbuntuDupQuestions,map,0.6615532012360636
|
29 |
+
sentence-t5-xxl,MindSmallReranking,map,0.3059647884122016
|
30 |
+
sentence-t5-xxl,SciDocsRR,map,0.7609412848204858
|
31 |
+
sentence-t5-xxl,StackOverflowDupQuestions,map,0.5285484177221984
|
32 |
+
sentence-t5-xxl,ArguAna,ndcg_at_10,0.39847
|
33 |
+
sentence-t5-xxl,ClimateFEVER,ndcg_at_10,
|
34 |
+
sentence-t5-xxl,CQADupstackRetrieval,ndcg_at_10,0.44654833333333344
|
35 |
+
sentence-t5-xxl,DBPedia,ndcg_at_10,
|
36 |
+
sentence-t5-xxl,FEVER,ndcg_at_10,
|
37 |
+
sentence-t5-xxl,FiQA2018,ndcg_at_10,0.46677
|
38 |
+
sentence-t5-xxl,HotpotQA,ndcg_at_10,
|
39 |
+
sentence-t5-xxl,MSMARCO,ndcg_at_10,
|
40 |
+
sentence-t5-xxl,NFCorpus,ndcg_at_10,0.35077
|
41 |
+
sentence-t5-xxl,NQ,ndcg_at_10,0.5287
|
42 |
+
sentence-t5-xxl,QuoraRetrieval,ndcg_at_10,0.85959
|
43 |
+
sentence-t5-xxl,SCIDOCS,ndcg_at_10,0.17173
|
44 |
+
sentence-t5-xxl,SciFact,ndcg_at_10,0.5538
|
45 |
+
sentence-t5-xxl,Touche2020,ndcg_at_10,0.21647
|
46 |
+
sentence-t5-xxl,TRECCOVID,ndcg_at_10,0.59481
|
47 |
+
sentence-t5-xxl,BIOSSES,cosine_spearman,0.8042954348508337
|
48 |
+
sentence-t5-xxl,SICK-R,cosine_spearman,0.8042954348508337
|
49 |
+
sentence-t5-xxl,STS12,cosine_spearman,0.7884633095838921
|
50 |
+
sentence-t5-xxl,STS13,cosine_spearman,0.8893834486158684
|
51 |
+
sentence-t5-xxl,STS14,cosine_spearman,0.8485684801116044
|
52 |
+
sentence-t5-xxl,STS15,cosine_spearman,0.8931982869751905
|
53 |
+
sentence-t5-xxl,STS16,cosine_spearman,0.846657396318755
|
54 |
+
sentence-t5-xxl,STS17,cosine_spearman,0.8946272351246337
|
55 |
+
sentence-t5-xxl,STS22,cosine_spearman,0.6532938965959216
|
56 |
+
sentence-t5-xxl,STSBenchmark,cosine_spearman,0.8401139141474803
|
57 |
+
sentence-t5-xxl,SummEval,cosine_spearman,0.30077056036328337
|
sgpt-bloom-7b1-msmarco_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,
|
3 |
-
AmazonPolarityClassification,accuracy,
|
4 |
-
AmazonReviewsClassification,accuracy,
|
5 |
-
Banking77Classification,accuracy,
|
6 |
-
EmotionClassification,accuracy,
|
7 |
-
ImdbClassification,accuracy,
|
8 |
-
MassiveIntentClassification,accuracy,
|
9 |
-
MassiveScenarioClassification,accuracy,
|
10 |
-
MTOPDomainClassification,accuracy,
|
11 |
-
MTOPIntentClassification,accuracy,
|
12 |
-
ToxicConversationsClassification,accuracy,
|
13 |
-
TweetSentimentExtractionClassification,accuracy,
|
14 |
-
ArxivClusteringP2P,v_measure,
|
15 |
-
ArxivClusteringS2S,v_measure,
|
16 |
-
BiorxivClusteringP2P,v_measure,
|
17 |
-
BiorxivClusteringS2S,v_measure,
|
18 |
-
MedrxivClusteringP2P,v_measure,
|
19 |
-
MedrxivClusteringS2S,v_measure,
|
20 |
-
RedditClustering,v_measure,
|
21 |
-
RedditClusteringP2P,v_measure,
|
22 |
-
StackExchangeClustering,v_measure,
|
23 |
-
StackExchangeClusteringP2P,v_measure,
|
24 |
-
TwentyNewsgroupsClustering,v_measure,
|
25 |
-
SprintDuplicateQuestions,ap,
|
26 |
-
TwitterSemEval2015,ap,
|
27 |
-
TwitterURLCorpus,ap,
|
28 |
-
AskUbuntuDupQuestions,map,
|
29 |
-
MindSmallReranking,map,
|
30 |
-
SciDocsRR,map,
|
31 |
-
StackOverflowDupQuestions,map,
|
32 |
-
ArguAna,ndcg_at_10,
|
33 |
-
ClimateFEVER,ndcg_at_10,
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,
|
35 |
-
DBPedia,ndcg_at_10,
|
36 |
-
FEVER,ndcg_at_10,
|
37 |
-
FiQA2018,ndcg_at_10,
|
38 |
-
HotpotQA,ndcg_at_10,
|
39 |
-
MSMARCO,ndcg_at_10,
|
40 |
-
NFCorpus,ndcg_at_10,
|
41 |
-
NQ,ndcg_at_10,
|
42 |
-
QuoraRetrieval,ndcg_at_10,
|
43 |
-
SCIDOCS,ndcg_at_10,
|
44 |
-
SciFact,ndcg_at_10,
|
45 |
-
Touche2020,ndcg_at_10,
|
46 |
-
TRECCOVID,ndcg_at_10,
|
47 |
-
BIOSSES,cosine_spearman,
|
48 |
-
SICK-R,cosine_spearman,
|
49 |
-
STS12,cosine_spearman,
|
50 |
-
STS13,cosine_spearman,
|
51 |
-
STS14,cosine_spearman,
|
52 |
-
STS15,cosine_spearman,
|
53 |
-
STS16,cosine_spearman,
|
54 |
-
STS17,cosine_spearman,
|
55 |
-
STS22,cosine_spearman,
|
56 |
-
STSBenchmark,cosine_spearman,
|
57 |
-
SummEval,cosine_spearman,
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
sgpt-bloom-7b1-msmarco,AmazonCounterfactualClassification,accuracy,
|
3 |
+
sgpt-bloom-7b1-msmarco,AmazonPolarityClassification,accuracy,
|
4 |
+
sgpt-bloom-7b1-msmarco,AmazonReviewsClassification,accuracy,
|
5 |
+
sgpt-bloom-7b1-msmarco,Banking77Classification,accuracy,
|
6 |
+
sgpt-bloom-7b1-msmarco,EmotionClassification,accuracy,
|
7 |
+
sgpt-bloom-7b1-msmarco,ImdbClassification,accuracy,
|
8 |
+
sgpt-bloom-7b1-msmarco,MassiveIntentClassification,accuracy,
|
9 |
+
sgpt-bloom-7b1-msmarco,MassiveScenarioClassification,accuracy,
|
10 |
+
sgpt-bloom-7b1-msmarco,MTOPDomainClassification,accuracy,
|
11 |
+
sgpt-bloom-7b1-msmarco,MTOPIntentClassification,accuracy,
|
12 |
+
sgpt-bloom-7b1-msmarco,ToxicConversationsClassification,accuracy,
|
13 |
+
sgpt-bloom-7b1-msmarco,TweetSentimentExtractionClassification,accuracy,
|
14 |
+
sgpt-bloom-7b1-msmarco,ArxivClusteringP2P,v_measure,
|
15 |
+
sgpt-bloom-7b1-msmarco,ArxivClusteringS2S,v_measure,
|
16 |
+
sgpt-bloom-7b1-msmarco,BiorxivClusteringP2P,v_measure,
|
17 |
+
sgpt-bloom-7b1-msmarco,BiorxivClusteringS2S,v_measure,
|
18 |
+
sgpt-bloom-7b1-msmarco,MedrxivClusteringP2P,v_measure,
|
19 |
+
sgpt-bloom-7b1-msmarco,MedrxivClusteringS2S,v_measure,
|
20 |
+
sgpt-bloom-7b1-msmarco,RedditClustering,v_measure,
|
21 |
+
sgpt-bloom-7b1-msmarco,RedditClusteringP2P,v_measure,
|
22 |
+
sgpt-bloom-7b1-msmarco,StackExchangeClustering,v_measure,
|
23 |
+
sgpt-bloom-7b1-msmarco,StackExchangeClusteringP2P,v_measure,
|
24 |
+
sgpt-bloom-7b1-msmarco,TwentyNewsgroupsClustering,v_measure,
|
25 |
+
sgpt-bloom-7b1-msmarco,SprintDuplicateQuestions,ap,
|
26 |
+
sgpt-bloom-7b1-msmarco,TwitterSemEval2015,ap,
|
27 |
+
sgpt-bloom-7b1-msmarco,TwitterURLCorpus,ap,
|
28 |
+
sgpt-bloom-7b1-msmarco,AskUbuntuDupQuestions,map,
|
29 |
+
sgpt-bloom-7b1-msmarco,MindSmallReranking,map,
|
30 |
+
sgpt-bloom-7b1-msmarco,SciDocsRR,map,
|
31 |
+
sgpt-bloom-7b1-msmarco,StackOverflowDupQuestions,map,
|
32 |
+
sgpt-bloom-7b1-msmarco,ArguAna,ndcg_at_10,
|
33 |
+
sgpt-bloom-7b1-msmarco,ClimateFEVER,ndcg_at_10,
|
34 |
+
sgpt-bloom-7b1-msmarco,CQADupstackRetrieval,ndcg_at_10,
|
35 |
+
sgpt-bloom-7b1-msmarco,DBPedia,ndcg_at_10,
|
36 |
+
sgpt-bloom-7b1-msmarco,FEVER,ndcg_at_10,
|
37 |
+
sgpt-bloom-7b1-msmarco,FiQA2018,ndcg_at_10,
|
38 |
+
sgpt-bloom-7b1-msmarco,HotpotQA,ndcg_at_10,
|
39 |
+
sgpt-bloom-7b1-msmarco,MSMARCO,ndcg_at_10,
|
40 |
+
sgpt-bloom-7b1-msmarco,NFCorpus,ndcg_at_10,
|
41 |
+
sgpt-bloom-7b1-msmarco,NQ,ndcg_at_10,
|
42 |
+
sgpt-bloom-7b1-msmarco,QuoraRetrieval,ndcg_at_10,
|
43 |
+
sgpt-bloom-7b1-msmarco,SCIDOCS,ndcg_at_10,
|
44 |
+
sgpt-bloom-7b1-msmarco,SciFact,ndcg_at_10,
|
45 |
+
sgpt-bloom-7b1-msmarco,Touche2020,ndcg_at_10,
|
46 |
+
sgpt-bloom-7b1-msmarco,TRECCOVID,ndcg_at_10,
|
47 |
+
sgpt-bloom-7b1-msmarco,BIOSSES,cosine_spearman,
|
48 |
+
sgpt-bloom-7b1-msmarco,SICK-R,cosine_spearman,
|
49 |
+
sgpt-bloom-7b1-msmarco,STS12,cosine_spearman,
|
50 |
+
sgpt-bloom-7b1-msmarco,STS13,cosine_spearman,
|
51 |
+
sgpt-bloom-7b1-msmarco,STS14,cosine_spearman,
|
52 |
+
sgpt-bloom-7b1-msmarco,STS15,cosine_spearman,
|
53 |
+
sgpt-bloom-7b1-msmarco,STS16,cosine_spearman,
|
54 |
+
sgpt-bloom-7b1-msmarco,STS17,cosine_spearman,
|
55 |
+
sgpt-bloom-7b1-msmarco,STS22,cosine_spearman,
|
56 |
+
sgpt-bloom-7b1-msmarco,STSBenchmark,cosine_spearman,
|
57 |
+
sgpt-bloom-7b1-msmarco,SummEval,cosine_spearman,
|
sup-simcse-bert-base-uncased_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.7574626865671641
|
3 |
-
AmazonPolarityClassification,accuracy,0.8247332500000001
|
4 |
-
AmazonReviewsClassification,accuracy,0.39598000000000005
|
5 |
-
Banking77Classification,accuracy,0.7575974025974027
|
6 |
-
EmotionClassification,accuracy,0.4481
|
7 |
-
ImdbClassification,accuracy,0.7353120000000001
|
8 |
-
MassiveIntentClassification,accuracy,0.6595158036314727
|
9 |
-
MassiveScenarioClassification,accuracy,0.7078345662407532
|
10 |
-
MTOPDomainClassification,accuracy,0.8428636570907433
|
11 |
-
MTOPIntentClassification,accuracy,0.6314409484724123
|
12 |
-
ToxicConversationsClassification,accuracy,0.720444
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5973401245048104
|
14 |
-
ArxivClusteringP2P,v_measure,0.35178847292844123
|
15 |
-
ArxivClusteringS2S,v_measure,0.2753955423760812
|
16 |
-
BiorxivClusteringP2P,v_measure,0.3015030662227709
|
17 |
-
BiorxivClusteringS2S,v_measure,0.24667587465067067
|
18 |
-
MedrxivClusteringP2P,v_measure,0.2624813884919727
|
19 |
-
MedrxivClusteringS2S,v_measure,0.24119581645899563
|
20 |
-
RedditClustering,v_measure,0.4022796798210619
|
21 |
-
RedditClusteringP2P,v_measure,0.4773761279854133
|
22 |
-
StackExchangeClustering,v_measure,0.4755177976262692
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.29453502158512795
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.3486285218255636
|
25 |
-
SprintDuplicateQuestions,ap,0.6938793582706408
|
26 |
-
TwitterSemEval2015,ap,0.6775361224485238
|
27 |
-
TwitterURLCorpus,ap,0.8388869258620997
|
28 |
-
AskUbuntuDupQuestions,map,0.5180158192605682
|
29 |
-
MindSmallReranking,map,0.2929834424526365
|
30 |
-
SciDocsRR,map,0.7013643643529786
|
31 |
-
StackOverflowDupQuestions,map,0.3890480943394549
|
32 |
-
ArguAna,ndcg_at_10,0.38331
|
33 |
-
ClimateFEVER,ndcg_at_10,0.11985
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.1449675
|
35 |
-
DBPedia,ndcg_at_10,0.19729
|
36 |
-
FEVER,ndcg_at_10,0.20411
|
37 |
-
FiQA2018,ndcg_at_10,0.1041
|
38 |
-
HotpotQA,ndcg_at_10,0.22895
|
39 |
-
MSMARCO,ndcg_at_10,0.25193
|
40 |
-
NFCorpus,ndcg_at_10,0.12418
|
41 |
-
NQ,ndcg_at_10,0.1608
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.79621
|
43 |
-
SCIDOCS,ndcg_at_10,0.07534
|
44 |
-
SciFact,ndcg_at_10,0.29593
|
45 |
-
Touche2020,ndcg_at_10,0.09895
|
46 |
-
TRECCOVID,ndcg_at_10,0.22929
|
47 |
-
BIOSSES,cosine_spearman,0.6837806406467221
|
48 |
-
SICK-R,cosine_spearman,0.8077099458895838
|
49 |
-
STS12,cosine_spearman,0.7529685124437535
|
50 |
-
STS13,cosine_spearman,0.8466699046070492
|
51 |
-
STS14,cosine_spearman,0.8018946888447316
|
52 |
-
STS15,cosine_spearman,0.8539881217395949
|
53 |
-
STS16,cosine_spearman,0.8081920987955081
|
54 |
-
STS17,cosine_spearman,0.8943673830561076
|
55 |
-
STS22,cosine_spearman,0.6195505621412989
|
56 |
-
STSBenchmark,cosine_spearman,0.8424697229482505
|
57 |
-
SummEval,cosine_spearman,0.23308772096174835
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
sup-simcse-bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7574626865671641
|
3 |
+
sup-simcse-bert-base-uncased,AmazonPolarityClassification,accuracy,0.8247332500000001
|
4 |
+
sup-simcse-bert-base-uncased,AmazonReviewsClassification,accuracy,0.39598000000000005
|
5 |
+
sup-simcse-bert-base-uncased,Banking77Classification,accuracy,0.7575974025974027
|
6 |
+
sup-simcse-bert-base-uncased,EmotionClassification,accuracy,0.4481
|
7 |
+
sup-simcse-bert-base-uncased,ImdbClassification,accuracy,0.7353120000000001
|
8 |
+
sup-simcse-bert-base-uncased,MassiveIntentClassification,accuracy,0.6595158036314727
|
9 |
+
sup-simcse-bert-base-uncased,MassiveScenarioClassification,accuracy,0.7078345662407532
|
10 |
+
sup-simcse-bert-base-uncased,MTOPDomainClassification,accuracy,0.8428636570907433
|
11 |
+
sup-simcse-bert-base-uncased,MTOPIntentClassification,accuracy,0.6314409484724123
|
12 |
+
sup-simcse-bert-base-uncased,ToxicConversationsClassification,accuracy,0.720444
|
13 |
+
sup-simcse-bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5973401245048104
|
14 |
+
sup-simcse-bert-base-uncased,ArxivClusteringP2P,v_measure,0.35178847292844123
|
15 |
+
sup-simcse-bert-base-uncased,ArxivClusteringS2S,v_measure,0.2753955423760812
|
16 |
+
sup-simcse-bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3015030662227709
|
17 |
+
sup-simcse-bert-base-uncased,BiorxivClusteringS2S,v_measure,0.24667587465067067
|
18 |
+
sup-simcse-bert-base-uncased,MedrxivClusteringP2P,v_measure,0.2624813884919727
|
19 |
+
sup-simcse-bert-base-uncased,MedrxivClusteringS2S,v_measure,0.24119581645899563
|
20 |
+
sup-simcse-bert-base-uncased,RedditClustering,v_measure,0.4022796798210619
|
21 |
+
sup-simcse-bert-base-uncased,RedditClusteringP2P,v_measure,0.4773761279854133
|
22 |
+
sup-simcse-bert-base-uncased,StackExchangeClustering,v_measure,0.4755177976262692
|
23 |
+
sup-simcse-bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.29453502158512795
|
24 |
+
sup-simcse-bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.3486285218255636
|
25 |
+
sup-simcse-bert-base-uncased,SprintDuplicateQuestions,ap,0.6938793582706408
|
26 |
+
sup-simcse-bert-base-uncased,TwitterSemEval2015,ap,0.6775361224485238
|
27 |
+
sup-simcse-bert-base-uncased,TwitterURLCorpus,ap,0.8388869258620997
|
28 |
+
sup-simcse-bert-base-uncased,AskUbuntuDupQuestions,map,0.5180158192605682
|
29 |
+
sup-simcse-bert-base-uncased,MindSmallReranking,map,0.2929834424526365
|
30 |
+
sup-simcse-bert-base-uncased,SciDocsRR,map,0.7013643643529786
|
31 |
+
sup-simcse-bert-base-uncased,StackOverflowDupQuestions,map,0.3890480943394549
|
32 |
+
sup-simcse-bert-base-uncased,ArguAna,ndcg_at_10,0.38331
|
33 |
+
sup-simcse-bert-base-uncased,ClimateFEVER,ndcg_at_10,0.11985
|
34 |
+
sup-simcse-bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.1449675
|
35 |
+
sup-simcse-bert-base-uncased,DBPedia,ndcg_at_10,0.19729
|
36 |
+
sup-simcse-bert-base-uncased,FEVER,ndcg_at_10,0.20411
|
37 |
+
sup-simcse-bert-base-uncased,FiQA2018,ndcg_at_10,0.1041
|
38 |
+
sup-simcse-bert-base-uncased,HotpotQA,ndcg_at_10,0.22895
|
39 |
+
sup-simcse-bert-base-uncased,MSMARCO,ndcg_at_10,0.25193
|
40 |
+
sup-simcse-bert-base-uncased,NFCorpus,ndcg_at_10,0.12418
|
41 |
+
sup-simcse-bert-base-uncased,NQ,ndcg_at_10,0.1608
|
42 |
+
sup-simcse-bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.79621
|
43 |
+
sup-simcse-bert-base-uncased,SCIDOCS,ndcg_at_10,0.07534
|
44 |
+
sup-simcse-bert-base-uncased,SciFact,ndcg_at_10,0.29593
|
45 |
+
sup-simcse-bert-base-uncased,Touche2020,ndcg_at_10,0.09895
|
46 |
+
sup-simcse-bert-base-uncased,TRECCOVID,ndcg_at_10,0.22929
|
47 |
+
sup-simcse-bert-base-uncased,BIOSSES,cosine_spearman,0.6837806406467221
|
48 |
+
sup-simcse-bert-base-uncased,SICK-R,cosine_spearman,0.8077099458895838
|
49 |
+
sup-simcse-bert-base-uncased,STS12,cosine_spearman,0.7529685124437535
|
50 |
+
sup-simcse-bert-base-uncased,STS13,cosine_spearman,0.8466699046070492
|
51 |
+
sup-simcse-bert-base-uncased,STS14,cosine_spearman,0.8018946888447316
|
52 |
+
sup-simcse-bert-base-uncased,STS15,cosine_spearman,0.8539881217395949
|
53 |
+
sup-simcse-bert-base-uncased,STS16,cosine_spearman,0.8081920987955081
|
54 |
+
sup-simcse-bert-base-uncased,STS17,cosine_spearman,0.8943673830561076
|
55 |
+
sup-simcse-bert-base-uncased,STS22,cosine_spearman,0.6195505621412989
|
56 |
+
sup-simcse-bert-base-uncased,STSBenchmark,cosine_spearman,0.8424697229482505
|
57 |
+
sup-simcse-bert-base-uncased,SummEval,cosine_spearman,0.23308772096174835
|
unsup-simcse-bert-base-uncased_results.csv
CHANGED
@@ -1,57 +1,57 @@
|
|
1 |
-
dataset,metric,value
|
2 |
-
AmazonCounterfactualClassification,accuracy,0.6708955223880597
|
3 |
-
AmazonPolarityClassification,accuracy,0.7447925
|
4 |
-
AmazonReviewsClassification,accuracy,0.33846000000000004
|
5 |
-
Banking77Classification,accuracy,0.735487012987013
|
6 |
-
EmotionClassification,accuracy,0.42219999999999996
|
7 |
-
ImdbClassification,accuracy,0.696284
|
8 |
-
MassiveIntentClassification,accuracy,0.5983860121049093
|
9 |
-
MassiveScenarioClassification,accuracy,0.662542030934768
|
10 |
-
MTOPDomainClassification,accuracy,0.817122663018696
|
11 |
-
MTOPIntentClassification,accuracy,0.5922708618331054
|
12 |
-
ToxicConversationsClassification,accuracy,0.688172
|
13 |
-
TweetSentimentExtractionClassification,accuracy,0.5335880022637238
|
14 |
-
ArxivClusteringP2P,v_measure,0.3260699748580497
|
15 |
-
ArxivClusteringS2S,v_measure,0.2468226709235907
|
16 |
-
BiorxivClusteringP2P,v_measure,0.24902697594424988
|
17 |
-
BiorxivClusteringS2S,v_measure,0.1955213639220657
|
18 |
-
MedrxivClusteringP2P,v_measure,0.23602716067058646
|
19 |
-
MedrxivClusteringS2S,v_measure,0.21973507645898568
|
20 |
-
RedditClustering,v_measure,0.32177188380770716
|
21 |
-
RedditClusteringP2P,v_measure,0.4513625539595167
|
22 |
-
StackExchangeClustering,v_measure,0.43070159841748956
|
23 |
-
StackExchangeClusteringP2P,v_measure,0.2850446706405736
|
24 |
-
TwentyNewsgroupsClustering,v_measure,0.23205752453237888
|
25 |
-
SprintDuplicateQuestions,ap,0.6940657828591321
|
26 |
-
TwitterSemEval2015,ap,0.6020770691133284
|
27 |
-
TwitterURLCorpus,ap,0.8136867122131795
|
28 |
-
AskUbuntuDupQuestions,map,0.5156997143788069
|
29 |
-
MindSmallReranking,map,0.2861705451549499
|
30 |
-
SciDocsRR,map,0.6633446765426328
|
31 |
-
StackOverflowDupQuestions,map,0.39352125808339417
|
32 |
-
ArguAna,ndcg_at_10,0.38343
|
33 |
-
ClimateFEVER,ndcg_at_10,0.11799
|
34 |
-
CQADupstackRetrieval,ndcg_at_10,0.13222750000000003
|
35 |
-
DBPedia,ndcg_at_10,0.15041
|
36 |
-
FEVER,ndcg_at_10,0.21055
|
37 |
-
FiQA2018,ndcg_at_10,0.09836
|
38 |
-
HotpotQA,ndcg_at_10,0.19752
|
39 |
-
MSMARCO,ndcg_at_10,0.22566
|
40 |
-
NFCorpus,ndcg_at_10,0.09879
|
41 |
-
NQ,ndcg_at_10,0.11692
|
42 |
-
QuoraRetrieval,ndcg_at_10,0.78031
|
43 |
-
SCIDOCS,ndcg_at_10,0.05496
|
44 |
-
SciFact,ndcg_at_10,0.25716
|
45 |
-
Touche2020,ndcg_at_10,0.08897
|
46 |
-
TRECCOVID,ndcg_at_10,0.26196
|
47 |
-
BIOSSES,cosine_spearman,0.7231128621592763
|
48 |
-
SICK-R,cosine_spearman,0.7224371728496006
|
49 |
-
STS12,cosine_spearman,0.6604669141177121
|
50 |
-
STS13,cosine_spearman,0.8148757487870733
|
51 |
-
STS14,cosine_spearman,0.7360735790527075
|
52 |
-
STS15,cosine_spearman,0.7972498368805553
|
53 |
-
STS16,cosine_spearman,0.7812450491845039
|
54 |
-
STS17,cosine_spearman,0.8357872706551984
|
55 |
-
STS22,cosine_spearman,0.5965048713602169
|
56 |
-
STSBenchmark,cosine_spearman,0.7651512563863317
|
57 |
-
SummEval,cosine_spearman,0.3114893410144058
|
|
|
1 |
+
model,dataset,metric,value
|
2 |
+
unsup-simcse-bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.6708955223880597
|
3 |
+
unsup-simcse-bert-base-uncased,AmazonPolarityClassification,accuracy,0.7447925
|
4 |
+
unsup-simcse-bert-base-uncased,AmazonReviewsClassification,accuracy,0.33846000000000004
|
5 |
+
unsup-simcse-bert-base-uncased,Banking77Classification,accuracy,0.735487012987013
|
6 |
+
unsup-simcse-bert-base-uncased,EmotionClassification,accuracy,0.42219999999999996
|
7 |
+
unsup-simcse-bert-base-uncased,ImdbClassification,accuracy,0.696284
|
8 |
+
unsup-simcse-bert-base-uncased,MassiveIntentClassification,accuracy,0.5983860121049093
|
9 |
+
unsup-simcse-bert-base-uncased,MassiveScenarioClassification,accuracy,0.662542030934768
|
10 |
+
unsup-simcse-bert-base-uncased,MTOPDomainClassification,accuracy,0.817122663018696
|
11 |
+
unsup-simcse-bert-base-uncased,MTOPIntentClassification,accuracy,0.5922708618331054
|
12 |
+
unsup-simcse-bert-base-uncased,ToxicConversationsClassification,accuracy,0.688172
|
13 |
+
unsup-simcse-bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5335880022637238
|
14 |
+
unsup-simcse-bert-base-uncased,ArxivClusteringP2P,v_measure,0.3260699748580497
|
15 |
+
unsup-simcse-bert-base-uncased,ArxivClusteringS2S,v_measure,0.2468226709235907
|
16 |
+
unsup-simcse-bert-base-uncased,BiorxivClusteringP2P,v_measure,0.24902697594424988
|
17 |
+
unsup-simcse-bert-base-uncased,BiorxivClusteringS2S,v_measure,0.1955213639220657
|
18 |
+
unsup-simcse-bert-base-uncased,MedrxivClusteringP2P,v_measure,0.23602716067058646
|
19 |
+
unsup-simcse-bert-base-uncased,MedrxivClusteringS2S,v_measure,0.21973507645898568
|
20 |
+
unsup-simcse-bert-base-uncased,RedditClustering,v_measure,0.32177188380770716
|
21 |
+
unsup-simcse-bert-base-uncased,RedditClusteringP2P,v_measure,0.4513625539595167
|
22 |
+
unsup-simcse-bert-base-uncased,StackExchangeClustering,v_measure,0.43070159841748956
|
23 |
+
unsup-simcse-bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2850446706405736
|
24 |
+
unsup-simcse-bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23205752453237888
|
25 |
+
unsup-simcse-bert-base-uncased,SprintDuplicateQuestions,ap,0.6940657828591321
|
26 |
+
unsup-simcse-bert-base-uncased,TwitterSemEval2015,ap,0.6020770691133284
|
27 |
+
unsup-simcse-bert-base-uncased,TwitterURLCorpus,ap,0.8136867122131795
|
28 |
+
unsup-simcse-bert-base-uncased,AskUbuntuDupQuestions,map,0.5156997143788069
|
29 |
+
unsup-simcse-bert-base-uncased,MindSmallReranking,map,0.2861705451549499
|
30 |
+
unsup-simcse-bert-base-uncased,SciDocsRR,map,0.6633446765426328
|
31 |
+
unsup-simcse-bert-base-uncased,StackOverflowDupQuestions,map,0.39352125808339417
|
32 |
+
unsup-simcse-bert-base-uncased,ArguAna,ndcg_at_10,0.38343
|
33 |
+
unsup-simcse-bert-base-uncased,ClimateFEVER,ndcg_at_10,0.11799
|
34 |
+
unsup-simcse-bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.13222750000000003
|
35 |
+
unsup-simcse-bert-base-uncased,DBPedia,ndcg_at_10,0.15041
|
36 |
+
unsup-simcse-bert-base-uncased,FEVER,ndcg_at_10,0.21055
|
37 |
+
unsup-simcse-bert-base-uncased,FiQA2018,ndcg_at_10,0.09836
|
38 |
+
unsup-simcse-bert-base-uncased,HotpotQA,ndcg_at_10,0.19752
|
39 |
+
unsup-simcse-bert-base-uncased,MSMARCO,ndcg_at_10,0.22566
|
40 |
+
unsup-simcse-bert-base-uncased,NFCorpus,ndcg_at_10,0.09879
|
41 |
+
unsup-simcse-bert-base-uncased,NQ,ndcg_at_10,0.11692
|
42 |
+
unsup-simcse-bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.78031
|
43 |
+
unsup-simcse-bert-base-uncased,SCIDOCS,ndcg_at_10,0.05496
|
44 |
+
unsup-simcse-bert-base-uncased,SciFact,ndcg_at_10,0.25716
|
45 |
+
unsup-simcse-bert-base-uncased,Touche2020,ndcg_at_10,0.08897
|
46 |
+
unsup-simcse-bert-base-uncased,TRECCOVID,ndcg_at_10,0.26196
|
47 |
+
unsup-simcse-bert-base-uncased,BIOSSES,cosine_spearman,0.7231128621592763
|
48 |
+
unsup-simcse-bert-base-uncased,SICK-R,cosine_spearman,0.7224371728496006
|
49 |
+
unsup-simcse-bert-base-uncased,STS12,cosine_spearman,0.6604669141177121
|
50 |
+
unsup-simcse-bert-base-uncased,STS13,cosine_spearman,0.8148757487870733
|
51 |
+
unsup-simcse-bert-base-uncased,STS14,cosine_spearman,0.7360735790527075
|
52 |
+
unsup-simcse-bert-base-uncased,STS15,cosine_spearman,0.7972498368805553
|
53 |
+
unsup-simcse-bert-base-uncased,STS16,cosine_spearman,0.7812450491845039
|
54 |
+
unsup-simcse-bert-base-uncased,STS17,cosine_spearman,0.8357872706551984
|
55 |
+
unsup-simcse-bert-base-uncased,STS22,cosine_spearman,0.5965048713602169
|
56 |
+
unsup-simcse-bert-base-uncased,STSBenchmark,cosine_spearman,0.7651512563863317
|
57 |
+
unsup-simcse-bert-base-uncased,SummEval,cosine_spearman,0.3114893410144058
|