tomaarsen's picture
tomaarsen HF staff
Upload all_results.json with huggingface_hub
c4cce95
{
"epoch": 3.0,
"test_art-broadcastprogram": {
"f1": 0.6051364365971107,
"number": 603,
"precision": 0.5863141524105754,
"recall": 0.6252072968490879
},
"test_art-film": {
"f1": 0.7647457627118643,
"number": 750,
"precision": 0.7779310344827586,
"recall": 0.752
},
"test_art-music": {
"f1": 0.7786106946526736,
"number": 1029,
"precision": 0.801440329218107,
"recall": 0.7570456754130224
},
"test_art-other": {
"f1": 0.3649193548387097,
"number": 562,
"precision": 0.42093023255813955,
"recall": 0.3220640569395018
},
"test_art-painting": {
"f1": 0.628099173553719,
"number": 57,
"precision": 0.59375,
"recall": 0.6666666666666666
},
"test_art-writtenart": {
"f1": 0.662753468516542,
"number": 968,
"precision": 0.6854304635761589,
"recall": 0.6415289256198347
},
"test_building-airport": {
"f1": 0.821917808219178,
"number": 364,
"precision": 0.819672131147541,
"recall": 0.8241758241758241
},
"test_building-hospital": {
"f1": 0.767052767052767,
"number": 364,
"precision": 0.7215496368038741,
"recall": 0.8186813186813187
},
"test_building-hotel": {
"f1": 0.7065637065637066,
"number": 265,
"precision": 0.7233201581027668,
"recall": 0.690566037735849
},
"test_building-library": {
"f1": 0.7424460431654676,
"number": 355,
"precision": 0.7588235294117647,
"recall": 0.7267605633802817
},
"test_building-other": {
"f1": 0.58483896307934,
"number": 2543,
"precision": 0.5841506473126716,
"recall": 0.5855289028706252
},
"test_building-restaurant": {
"f1": 0.5195402298850574,
"number": 232,
"precision": 0.5566502463054187,
"recall": 0.4870689655172414
},
"test_building-sportsfacility": {
"f1": 0.7052401746724892,
"number": 420,
"precision": 0.6512096774193549,
"recall": 0.7690476190476191
},
"test_building-theater": {
"f1": 0.7245762711864406,
"number": 455,
"precision": 0.6993865030674846,
"recall": 0.7516483516483516
},
"test_event-attack/battle/war/militaryconflict": {
"f1": 0.755868544600939,
"number": 1098,
"precision": 0.7800387596899225,
"recall": 0.7331511839708561
},
"test_event-disaster": {
"f1": 0.5505050505050505,
"number": 207,
"precision": 0.5767195767195767,
"recall": 0.5265700483091788
},
"test_event-election": {
"f1": 0.2096069868995633,
"number": 182,
"precision": 0.5106382978723404,
"recall": 0.13186813186813187
},
"test_event-other": {
"f1": 0.4504391468005019,
"number": 866,
"precision": 0.49313186813186816,
"recall": 0.41454965357967666
},
"test_event-protest": {
"f1": 0.39999999999999997,
"number": 166,
"precision": 0.3711340206185567,
"recall": 0.43373493975903615
},
"test_event-sportsevent": {
"f1": 0.6155810983397191,
"number": 1566,
"precision": 0.6155810983397191,
"recall": 0.6155810983397191
},
"test_location-GPE": {
"f1": 0.8338255420298207,
"number": 20409,
"precision": 0.8175141242937853,
"recall": 0.8508011171541967
},
"test_location-bodiesofwater": {
"f1": 0.7456066945606695,
"number": 1169,
"precision": 0.7297297297297297,
"recall": 0.7621899059024807
},
"test_location-island": {
"f1": 0.6995153473344103,
"number": 646,
"precision": 0.731418918918919,
"recall": 0.6702786377708978
},
"test_location-mountain": {
"f1": 0.7408513816280807,
"number": 681,
"precision": 0.7537993920972644,
"recall": 0.7283406754772394
},
"test_location-other": {
"f1": 0.3585464333781965,
"number": 2191,
"precision": 0.43700787401574803,
"recall": 0.3039707895937928
},
"test_location-park": {
"f1": 0.6969026548672567,
"number": 458,
"precision": 0.7062780269058296,
"recall": 0.6877729257641921
},
"test_location-road/railway/highway/transit": {
"f1": 0.7174418604651162,
"number": 1700,
"precision": 0.7091954022988506,
"recall": 0.7258823529411764
},
"test_loss": 0.022720418870449066,
"test_organization-company": {
"f1": 0.6927016645326505,
"number": 3896,
"precision": 0.6911088400613183,
"recall": 0.6943018480492813
},
"test_organization-education": {
"f1": 0.7885167464114833,
"number": 2067,
"precision": 0.7799337434926644,
"recall": 0.7972907595549105
},
"test_organization-government/governmentagency": {
"f1": 0.4941520467836257,
"number": 1511,
"precision": 0.5518367346938775,
"recall": 0.44738583719391134
},
"test_organization-media/newspaper": {
"f1": 0.6505271378367826,
"number": 1232,
"precision": 0.6267870579382995,
"recall": 0.6761363636363636
},
"test_organization-other": {
"f1": 0.5563115908024402,
"number": 4439,
"precision": 0.5804161566707466,
"recall": 0.5341293084027934
},
"test_organization-politicalparty": {
"f1": 0.6949458483754513,
"number": 1054,
"precision": 0.6626506024096386,
"recall": 0.7305502846299811
},
"test_organization-religion": {
"f1": 0.5933756166314307,
"number": 672,
"precision": 0.5635876840696118,
"recall": 0.6264880952380952
},
"test_organization-showorganization": {
"f1": 0.6054333764553688,
"number": 769,
"precision": 0.6023166023166023,
"recall": 0.6085825747724317
},
"test_organization-sportsleague": {
"f1": 0.6544831524842947,
"number": 882,
"precision": 0.6593785960874569,
"recall": 0.6496598639455783
},
"test_organization-sportsteam": {
"f1": 0.7517758484609314,
"number": 2473,
"precision": 0.7341040462427746,
"recall": 0.770319450060655
},
"test_other-astronomything": {
"f1": 0.8040057224606582,
"number": 678,
"precision": 0.7805555555555556,
"recall": 0.8289085545722714
},
"test_other-award": {
"f1": 0.6956521739130435,
"number": 919,
"precision": 0.7230046948356808,
"recall": 0.6702937976060935
},
"test_other-biologything": {
"f1": 0.6544157981349424,
"number": 1874,
"precision": 0.6732505643340858,
"recall": 0.6366061899679829
},
"test_other-chemicalthing": {
"f1": 0.5899352267065271,
"number": 1014,
"precision": 0.5961732124874118,
"recall": 0.5838264299802761
},
"test_other-currency": {
"f1": 0.746268656716418,
"number": 799,
"precision": 0.7134703196347032,
"recall": 0.7822277847309136
},
"test_other-disease": {
"f1": 0.6637390213299874,
"number": 749,
"precision": 0.6260355029585799,
"recall": 0.7062750333778371
},
"test_other-educationaldegree": {
"f1": 0.6016483516483516,
"number": 363,
"precision": 0.6,
"recall": 0.6033057851239669
},
"test_other-god": {
"f1": 0.7084639498432602,
"number": 635,
"precision": 0.7051482059282371,
"recall": 0.7118110236220473
},
"test_other-language": {
"f1": 0.736648250460405,
"number": 753,
"precision": 0.684931506849315,
"recall": 0.796812749003984
},
"test_other-law": {
"f1": 0.6828752642706131,
"number": 472,
"precision": 0.6814345991561181,
"recall": 0.684322033898305
},
"test_other-livingthing": {
"f1": 0.6191536748329621,
"number": 863,
"precision": 0.5959271168274384,
"recall": 0.6442641946697567
},
"test_other-medical": {
"f1": 0.5019710906701709,
"number": 397,
"precision": 0.5247252747252747,
"recall": 0.4811083123425693
},
"test_overall_accuracy": 0.9248186428918111,
"test_overall_f1": 0.7006507253689264,
"test_overall_precision": 0.7040676584045078,
"test_overall_recall": 0.6972667978051558,
"test_person-actor": {
"f1": 0.8146295717411691,
"number": 1637,
"precision": 0.8341869398207427,
"recall": 0.7959682345754429
},
"test_person-artist/author": {
"f1": 0.7260753818130867,
"number": 3463,
"precision": 0.7052259118127382,
"recall": 0.74819520646838
},
"test_person-athlete": {
"f1": 0.8462332301341589,
"number": 2884,
"precision": 0.8395904436860068,
"recall": 0.8529819694868238
},
"test_person-director": {
"f1": 0.7289048473967685,
"number": 554,
"precision": 0.725,
"recall": 0.7328519855595668
},
"test_person-other": {
"f1": 0.6767326159898183,
"number": 8767,
"precision": 0.6865829322690457,
"recall": 0.6671609444507813
},
"test_person-politician": {
"f1": 0.6835310537334263,
"number": 2859,
"precision": 0.6818656456665506,
"recall": 0.6852046169989506
},
"test_person-scholar": {
"f1": 0.5197740112994349,
"number": 743,
"precision": 0.5468053491827637,
"recall": 0.4952893674293405
},
"test_person-soldier": {
"f1": 0.5496987951807228,
"number": 647,
"precision": 0.5359765051395007,
"recall": 0.5641421947449768
},
"test_product-airplane": {
"f1": 0.6776859504132232,
"number": 792,
"precision": 0.6824583866837388,
"recall": 0.672979797979798
},
"test_product-car": {
"f1": 0.7109144542772862,
"number": 687,
"precision": 0.7204783258594918,
"recall": 0.7016011644832606
},
"test_product-food": {
"f1": 0.5696821515892421,
"number": 432,
"precision": 0.6036269430051814,
"recall": 0.5393518518518519
},
"test_product-game": {
"f1": 0.728249194414608,
"number": 493,
"precision": 0.773972602739726,
"recall": 0.6876267748478702
},
"test_product-other": {
"f1": 0.4614848379226211,
"number": 1608,
"precision": 0.5249801744647106,
"recall": 0.4116915422885572
},
"test_product-ship": {
"f1": 0.6772068511198946,
"number": 380,
"precision": 0.6781002638522428,
"recall": 0.6763157894736842
},
"test_product-software": {
"f1": 0.6651558073654391,
"number": 889,
"precision": 0.6700913242009132,
"recall": 0.6602924634420697
},
"test_product-train": {
"f1": 0.5984251968503936,
"number": 314,
"precision": 0.5919003115264797,
"recall": 0.6050955414012739
},
"test_product-weapon": {
"f1": 0.5921397379912663,
"number": 624,
"precision": 0.6506717850287908,
"recall": 0.5432692307692307
},
"test_runtime": 1023.4462,
"test_samples_per_second": 45.081,
"test_steps_per_second": 2.818
}