results / bert-base-uncased_results.csv
nouamanetazi's picture
nouamanetazi HF staff
add model column
b475468
raw history blame
No virus
3.48 kB
model,dataset,metric,value
bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7425373134328358
bert-base-uncased,AmazonPolarityClassification,accuracy,0.7132945
bert-base-uncased,AmazonReviewsClassification,accuracy,0.33564
bert-base-uncased,Banking77Classification,accuracy,0.6340584415584415
bert-base-uncased,EmotionClassification,accuracy,0.3528
bert-base-uncased,ImdbClassification,accuracy,0.653456
bert-base-uncased,MassiveIntentClassification,accuracy,0.5988231338264962
bert-base-uncased,MassiveScenarioClassification,accuracy,0.6427706792199059
bert-base-uncased,MTOPDomainClassification,accuracy,0.8262653898768809
bert-base-uncased,MTOPIntentClassification,accuracy,0.6813725490196079
bert-base-uncased,ToxicConversationsClassification,accuracy,0.6999679999999999
bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5180814940577249
bert-base-uncased,ArxivClusteringP2P,v_measure,0.3518932830729758
bert-base-uncased,ArxivClusteringS2S,v_measure,0.2750822951507033
bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3012278640553615
bert-base-uncased,BiorxivClusteringS2S,v_measure,0.2476601053047804
bert-base-uncased,MedrxivClusteringP2P,v_measure,0.26087688306606044
bert-base-uncased,MedrxivClusteringS2S,v_measure,0.23604914608225602
bert-base-uncased,RedditClustering,v_measure,0.27241718557705
bert-base-uncased,RedditClusteringP2P,v_measure,0.43323854460464056
bert-base-uncased,StackExchangeClustering,v_measure,0.4358261870547655
bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2654851675795123
bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23354321403189832
bert-base-uncased,SprintDuplicateQuestions,ap,0.36808703728970593
bert-base-uncased,TwitterSemEval2015,ap,0.5589788752976392
bert-base-uncased,TwitterURLCorpus,ap,0.7628732783437441
bert-base-uncased,AskUbuntuDupQuestions,map,0.4584088706528255
bert-base-uncased,MindSmallReranking,map,0.28366637355845425
bert-base-uncased,SciDocsRR,map,0.6493728987830145
bert-base-uncased,StackOverflowDupQuestions,map,0.34615478798860166
bert-base-uncased,ArguAna,ndcg_at_10,0.28294
bert-base-uncased,ClimateFEVER,ndcg_at_10,0.0541
bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
bert-base-uncased,DBPedia,ndcg_at_10,0.04132
bert-base-uncased,FEVER,ndcg_at_10,0.033
bert-base-uncased,FiQA2018,ndcg_at_10,0.02191
bert-base-uncased,HotpotQA,ndcg_at_10,0.0826
bert-base-uncased,MSMARCO,ndcg_at_10,0.06176
bert-base-uncased,NFCorpus,ndcg_at_10,0.04304
bert-base-uncased,NQ,ndcg_at_10,0.02615
bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.61029
bert-base-uncased,SCIDOCS,ndcg_at_10,0.02815
bert-base-uncased,SciFact,ndcg_at_10,0.13339
bert-base-uncased,Touche2020,ndcg_at_10,0.00967
bert-base-uncased,TRECCOVID,ndcg_at_10,0.14745
bert-base-uncased,BIOSSES,cosine_spearman,0.5469823428818151
bert-base-uncased,SICK-R,cosine_spearman,0.5864506948179484
bert-base-uncased,STS12,cosine_spearman,0.30871788407575457
bert-base-uncased,STS13,cosine_spearman,0.5989485045425808
bert-base-uncased,STS14,cosine_spearman,0.4772791352844941
bert-base-uncased,STS15,cosine_spearman,0.6028567383446486
bert-base-uncased,STS16,cosine_spearman,0.6373272693604403
bert-base-uncased,STS17,cosine_spearman,0.6410023907260639
bert-base-uncased,STS22,cosine_spearman,0.563667996535454
bert-base-uncased,STSBenchmark,cosine_spearman,0.4729108172174081
bert-base-uncased,SummEval,cosine_spearman,0.2981716355664815