|
{ |
|
"best_metric": 1.0657285451889038, |
|
"best_model_checkpoint": "actsa-distilbert/checkpoint-3246", |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 3246, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02310536044362292, |
|
"grad_norm": 2.318068265914917, |
|
"learning_rate": 3.846153846153846e-05, |
|
"loss": 1.0841, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.04621072088724584, |
|
"grad_norm": 3.620454788208008, |
|
"learning_rate": 7.692307692307693e-05, |
|
"loss": 1.0888, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06931608133086876, |
|
"grad_norm": 2.4006574153900146, |
|
"learning_rate": 0.0001153846153846154, |
|
"loss": 1.0742, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.09242144177449169, |
|
"grad_norm": 4.676277160644531, |
|
"learning_rate": 0.00015384615384615385, |
|
"loss": 1.0871, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.11552680221811461, |
|
"grad_norm": 6.01164436340332, |
|
"learning_rate": 0.00019230769230769233, |
|
"loss": 1.1146, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.13863216266173753, |
|
"grad_norm": 2.2937514781951904, |
|
"learning_rate": 0.0002307692307692308, |
|
"loss": 1.0578, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.16173752310536044, |
|
"grad_norm": 2.6013343334198, |
|
"learning_rate": 0.0002692307692307692, |
|
"loss": 1.0571, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.18484288354898337, |
|
"grad_norm": 3.571342945098877, |
|
"learning_rate": 0.0003076923076923077, |
|
"loss": 1.145, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.20794824399260628, |
|
"grad_norm": 2.3125054836273193, |
|
"learning_rate": 0.00034615384615384613, |
|
"loss": 1.0901, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.23105360443622922, |
|
"grad_norm": 5.343853950500488, |
|
"learning_rate": 0.00038461538461538467, |
|
"loss": 1.0198, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2541589648798521, |
|
"grad_norm": 4.148077487945557, |
|
"learning_rate": 0.0004230769230769231, |
|
"loss": 1.0066, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.27726432532347506, |
|
"grad_norm": 3.7557380199432373, |
|
"learning_rate": 0.0004615384615384616, |
|
"loss": 1.0734, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.300369685767098, |
|
"grad_norm": 3.485323190689087, |
|
"learning_rate": 0.0005, |
|
"loss": 1.0569, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.3234750462107209, |
|
"grad_norm": 2.63746976852417, |
|
"learning_rate": 0.0004957206436152003, |
|
"loss": 1.1081, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.3465804066543438, |
|
"grad_norm": 5.485255718231201, |
|
"learning_rate": 0.0004914412872304005, |
|
"loss": 1.1346, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.36968576709796674, |
|
"grad_norm": 2.920285940170288, |
|
"learning_rate": 0.00048716193084560086, |
|
"loss": 1.0998, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3927911275415896, |
|
"grad_norm": 2.696606397628784, |
|
"learning_rate": 0.0004828825744608011, |
|
"loss": 1.1216, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.41589648798521256, |
|
"grad_norm": 2.0741255283355713, |
|
"learning_rate": 0.0004786032180760014, |
|
"loss": 1.1092, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4390018484288355, |
|
"grad_norm": 2.102992534637451, |
|
"learning_rate": 0.0004743238616912017, |
|
"loss": 1.1172, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.46210720887245843, |
|
"grad_norm": 1.4443374872207642, |
|
"learning_rate": 0.0004700445053064019, |
|
"loss": 1.0929, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4852125693160813, |
|
"grad_norm": 2.7878777980804443, |
|
"learning_rate": 0.0004657651489216022, |
|
"loss": 1.125, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.5083179297597042, |
|
"grad_norm": 2.387270212173462, |
|
"learning_rate": 0.00046148579253680244, |
|
"loss": 1.1069, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5314232902033271, |
|
"grad_norm": 2.1369564533233643, |
|
"learning_rate": 0.00045720643615200274, |
|
"loss": 1.085, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.5545286506469501, |
|
"grad_norm": 2.407505989074707, |
|
"learning_rate": 0.00045292707976720304, |
|
"loss": 1.095, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.577634011090573, |
|
"grad_norm": 1.3333139419555664, |
|
"learning_rate": 0.0004486477233824033, |
|
"loss": 1.1035, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.600739371534196, |
|
"grad_norm": 1.891635537147522, |
|
"learning_rate": 0.0004443683669976036, |
|
"loss": 0.9419, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6238447319778189, |
|
"grad_norm": 2.536003589630127, |
|
"learning_rate": 0.00044008901061280383, |
|
"loss": 1.1685, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.6469500924214417, |
|
"grad_norm": 2.3016841411590576, |
|
"learning_rate": 0.00043580965422800413, |
|
"loss": 1.0619, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6700554528650647, |
|
"grad_norm": 1.4672470092773438, |
|
"learning_rate": 0.00043153029784320443, |
|
"loss": 1.1276, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.6931608133086876, |
|
"grad_norm": 2.2047739028930664, |
|
"learning_rate": 0.0004272509414584047, |
|
"loss": 1.0428, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.7162661737523105, |
|
"grad_norm": 4.394761562347412, |
|
"learning_rate": 0.000422971585073605, |
|
"loss": 1.082, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.7393715341959335, |
|
"grad_norm": 1.6812162399291992, |
|
"learning_rate": 0.00041869222868880517, |
|
"loss": 1.1325, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7624768946395564, |
|
"grad_norm": 2.4136714935302734, |
|
"learning_rate": 0.00041441287230400547, |
|
"loss": 1.0486, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.7855822550831792, |
|
"grad_norm": 1.2103418111801147, |
|
"learning_rate": 0.00041013351591920577, |
|
"loss": 1.093, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.8086876155268022, |
|
"grad_norm": 2.494950771331787, |
|
"learning_rate": 0.000405854159534406, |
|
"loss": 1.0638, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.8317929759704251, |
|
"grad_norm": 1.8322207927703857, |
|
"learning_rate": 0.0004015748031496063, |
|
"loss": 0.9711, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.8548983364140481, |
|
"grad_norm": 2.9408106803894043, |
|
"learning_rate": 0.00039729544676480656, |
|
"loss": 1.1413, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.878003696857671, |
|
"grad_norm": 1.0680229663848877, |
|
"learning_rate": 0.00039301609038000686, |
|
"loss": 1.0609, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.9011090573012939, |
|
"grad_norm": 1.952222466468811, |
|
"learning_rate": 0.00038873673399520716, |
|
"loss": 1.0954, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.9242144177449169, |
|
"grad_norm": 3.1969923973083496, |
|
"learning_rate": 0.0003844573776104074, |
|
"loss": 1.0895, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.9473197781885397, |
|
"grad_norm": 2.0018506050109863, |
|
"learning_rate": 0.0003801780212256077, |
|
"loss": 1.0697, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.9704251386321626, |
|
"grad_norm": 2.0683329105377197, |
|
"learning_rate": 0.00037589866484080795, |
|
"loss": 1.0999, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.9935304990757856, |
|
"grad_norm": 2.1113333702087402, |
|
"learning_rate": 0.00037161930845600825, |
|
"loss": 1.1089, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.4584103512014787, |
|
"eval_f1_macro": 0.2095479509928179, |
|
"eval_f1_micro": 0.4584103512014787, |
|
"eval_f1_weighted": 0.2881768494245037, |
|
"eval_loss": 1.0685529708862305, |
|
"eval_precision_macro": 0.1528034504004929, |
|
"eval_precision_micro": 0.4584103512014787, |
|
"eval_precision_weighted": 0.21014005008866307, |
|
"eval_recall_macro": 0.3333333333333333, |
|
"eval_recall_micro": 0.4584103512014787, |
|
"eval_recall_weighted": 0.4584103512014787, |
|
"eval_runtime": 103.9446, |
|
"eval_samples_per_second": 10.409, |
|
"eval_steps_per_second": 1.308, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 1.0166358595194085, |
|
"grad_norm": 2.078071117401123, |
|
"learning_rate": 0.0003673399520712085, |
|
"loss": 1.0538, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.0397412199630314, |
|
"grad_norm": 1.2932239770889282, |
|
"learning_rate": 0.00036306059568640874, |
|
"loss": 1.1108, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.0628465804066543, |
|
"grad_norm": 1.8026275634765625, |
|
"learning_rate": 0.00035878123930160904, |
|
"loss": 1.0782, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.0859519408502774, |
|
"grad_norm": 2.0047168731689453, |
|
"learning_rate": 0.0003545018829168093, |
|
"loss": 1.0754, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.1090573012939002, |
|
"grad_norm": 2.5215866565704346, |
|
"learning_rate": 0.0003502225265320096, |
|
"loss": 1.0822, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.1321626617375231, |
|
"grad_norm": 1.7528326511383057, |
|
"learning_rate": 0.0003459431701472099, |
|
"loss": 1.0956, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.155268022181146, |
|
"grad_norm": 2.345102548599243, |
|
"learning_rate": 0.00034166381376241013, |
|
"loss": 1.0284, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.1783733826247689, |
|
"grad_norm": 2.051119327545166, |
|
"learning_rate": 0.00033738445737761043, |
|
"loss": 1.0811, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.201478743068392, |
|
"grad_norm": 2.405728340148926, |
|
"learning_rate": 0.0003331051009928107, |
|
"loss": 1.0661, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.2245841035120149, |
|
"grad_norm": 2.118389368057251, |
|
"learning_rate": 0.000328825744608011, |
|
"loss": 1.0821, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.2476894639556377, |
|
"grad_norm": 1.383575677871704, |
|
"learning_rate": 0.0003245463882232113, |
|
"loss": 1.0988, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.2707948243992606, |
|
"grad_norm": 0.9262095093727112, |
|
"learning_rate": 0.00032026703183841147, |
|
"loss": 1.0855, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.2939001848428835, |
|
"grad_norm": 0.8497403860092163, |
|
"learning_rate": 0.00031598767545361177, |
|
"loss": 1.0734, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.3170055452865066, |
|
"grad_norm": 2.554072856903076, |
|
"learning_rate": 0.000311708319068812, |
|
"loss": 1.0495, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.3401109057301293, |
|
"grad_norm": 2.196169853210449, |
|
"learning_rate": 0.0003074289626840123, |
|
"loss": 1.0611, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.3632162661737524, |
|
"grad_norm": 1.9735369682312012, |
|
"learning_rate": 0.0003031496062992126, |
|
"loss": 1.0689, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.3863216266173752, |
|
"grad_norm": 1.28324556350708, |
|
"learning_rate": 0.00029887024991441286, |
|
"loss": 1.1195, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.4094269870609981, |
|
"grad_norm": 2.870333671569824, |
|
"learning_rate": 0.00029459089352961316, |
|
"loss": 1.0909, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.432532347504621, |
|
"grad_norm": 2.790924549102783, |
|
"learning_rate": 0.00029031153714481346, |
|
"loss": 1.1001, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.4556377079482439, |
|
"grad_norm": 1.2432293891906738, |
|
"learning_rate": 0.0002860321807600137, |
|
"loss": 1.124, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.478743068391867, |
|
"grad_norm": 0.8692905306816101, |
|
"learning_rate": 0.000281752824375214, |
|
"loss": 1.0717, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.5018484288354899, |
|
"grad_norm": 1.2673799991607666, |
|
"learning_rate": 0.00027747346799041425, |
|
"loss": 1.086, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.5249537892791127, |
|
"grad_norm": 1.9591180086135864, |
|
"learning_rate": 0.00027319411160561455, |
|
"loss": 1.0408, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.5480591497227358, |
|
"grad_norm": 1.8689022064208984, |
|
"learning_rate": 0.00026891475522081485, |
|
"loss": 1.0499, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.5711645101663585, |
|
"grad_norm": 2.0180625915527344, |
|
"learning_rate": 0.00026463539883601504, |
|
"loss": 1.0936, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.5942698706099816, |
|
"grad_norm": 0.9715362191200256, |
|
"learning_rate": 0.00026035604245121534, |
|
"loss": 1.0886, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.6173752310536045, |
|
"grad_norm": 1.5542079210281372, |
|
"learning_rate": 0.0002560766860664156, |
|
"loss": 1.0825, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.6404805914972274, |
|
"grad_norm": 2.8334264755249023, |
|
"learning_rate": 0.0002517973296816159, |
|
"loss": 1.0932, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.6635859519408502, |
|
"grad_norm": 0.5704463720321655, |
|
"learning_rate": 0.0002475179732968162, |
|
"loss": 1.1057, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.6866913123844731, |
|
"grad_norm": 2.452209949493408, |
|
"learning_rate": 0.00024323861691201643, |
|
"loss": 1.0569, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.7097966728280962, |
|
"grad_norm": 1.304701805114746, |
|
"learning_rate": 0.0002389592605272167, |
|
"loss": 1.0859, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.7329020332717189, |
|
"grad_norm": 1.5926436185836792, |
|
"learning_rate": 0.000234679904142417, |
|
"loss": 1.0832, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.756007393715342, |
|
"grad_norm": 2.683790445327759, |
|
"learning_rate": 0.00023040054775761728, |
|
"loss": 1.0539, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.7791127541589649, |
|
"grad_norm": 1.4463739395141602, |
|
"learning_rate": 0.00022612119137281752, |
|
"loss": 1.0358, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.8022181146025877, |
|
"grad_norm": 3.22567081451416, |
|
"learning_rate": 0.0002218418349880178, |
|
"loss": 0.9887, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.8253234750462108, |
|
"grad_norm": 0.9853553771972656, |
|
"learning_rate": 0.00021756247860321807, |
|
"loss": 1.0566, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.8484288354898335, |
|
"grad_norm": 1.0822279453277588, |
|
"learning_rate": 0.00021328312221841837, |
|
"loss": 1.0437, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.8715341959334566, |
|
"grad_norm": 1.9909262657165527, |
|
"learning_rate": 0.00020900376583361864, |
|
"loss": 1.0059, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.8946395563770795, |
|
"grad_norm": 1.8531982898712158, |
|
"learning_rate": 0.0002047244094488189, |
|
"loss": 1.0671, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.9177449168207024, |
|
"grad_norm": 1.0466595888137817, |
|
"learning_rate": 0.00020044505306401916, |
|
"loss": 1.0766, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 1.9408502772643255, |
|
"grad_norm": 0.8994802236557007, |
|
"learning_rate": 0.00019616569667921946, |
|
"loss": 0.9991, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.9639556377079481, |
|
"grad_norm": 1.929488182067871, |
|
"learning_rate": 0.00019188634029441973, |
|
"loss": 1.0905, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 1.9870609981515712, |
|
"grad_norm": 1.9165269136428833, |
|
"learning_rate": 0.00018760698390962, |
|
"loss": 1.039, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.4584103512014787, |
|
"eval_f1_macro": 0.2095479509928179, |
|
"eval_f1_micro": 0.4584103512014787, |
|
"eval_f1_weighted": 0.2881768494245037, |
|
"eval_loss": 1.0747383832931519, |
|
"eval_precision_macro": 0.1528034504004929, |
|
"eval_precision_micro": 0.4584103512014787, |
|
"eval_precision_weighted": 0.21014005008866307, |
|
"eval_recall_macro": 0.3333333333333333, |
|
"eval_recall_micro": 0.4584103512014787, |
|
"eval_recall_weighted": 0.4584103512014787, |
|
"eval_runtime": 105.3751, |
|
"eval_samples_per_second": 10.268, |
|
"eval_steps_per_second": 1.291, |
|
"step": 2164 |
|
}, |
|
{ |
|
"epoch": 2.010166358595194, |
|
"grad_norm": 4.0738606452941895, |
|
"learning_rate": 0.00018332762752482028, |
|
"loss": 1.0268, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 2.033271719038817, |
|
"grad_norm": 2.0423989295959473, |
|
"learning_rate": 0.00017904827114002055, |
|
"loss": 1.0547, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.05637707948244, |
|
"grad_norm": 2.3596909046173096, |
|
"learning_rate": 0.00017476891475522082, |
|
"loss": 1.0647, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 2.0794824399260627, |
|
"grad_norm": 1.9593594074249268, |
|
"learning_rate": 0.0001704895583704211, |
|
"loss": 1.1023, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.102587800369686, |
|
"grad_norm": 1.8953109979629517, |
|
"learning_rate": 0.00016621020198562137, |
|
"loss": 1.093, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 2.1256931608133085, |
|
"grad_norm": 1.6998157501220703, |
|
"learning_rate": 0.00016193084560082164, |
|
"loss": 1.0777, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.1487985212569316, |
|
"grad_norm": 1.0082741975784302, |
|
"learning_rate": 0.0001576514892160219, |
|
"loss": 1.0999, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 2.1719038817005547, |
|
"grad_norm": 1.887078046798706, |
|
"learning_rate": 0.0001533721328312222, |
|
"loss": 1.1067, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.1950092421441774, |
|
"grad_norm": 1.8659969568252563, |
|
"learning_rate": 0.00014909277644642246, |
|
"loss": 1.0769, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 2.2181146025878005, |
|
"grad_norm": 0.916480302810669, |
|
"learning_rate": 0.00014481342006162273, |
|
"loss": 1.048, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.241219963031423, |
|
"grad_norm": 1.9037351608276367, |
|
"learning_rate": 0.000140534063676823, |
|
"loss": 1.0509, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 2.2643253234750462, |
|
"grad_norm": 1.9867937564849854, |
|
"learning_rate": 0.00013625470729202328, |
|
"loss": 1.0609, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.287430683918669, |
|
"grad_norm": 2.044760227203369, |
|
"learning_rate": 0.00013197535090722358, |
|
"loss": 1.0351, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 2.310536044362292, |
|
"grad_norm": 2.0097403526306152, |
|
"learning_rate": 0.00012769599452242385, |
|
"loss": 1.0433, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.333641404805915, |
|
"grad_norm": 1.9191139936447144, |
|
"learning_rate": 0.0001234166381376241, |
|
"loss": 1.0219, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 2.3567467652495377, |
|
"grad_norm": 1.9857523441314697, |
|
"learning_rate": 0.00011913728175282437, |
|
"loss": 1.0892, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.379852125693161, |
|
"grad_norm": 2.7910494804382324, |
|
"learning_rate": 0.00011485792536802465, |
|
"loss": 1.1069, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 2.402957486136784, |
|
"grad_norm": 1.5788651704788208, |
|
"learning_rate": 0.00011057856898322493, |
|
"loss": 1.0685, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.4260628465804066, |
|
"grad_norm": 2.1729698181152344, |
|
"learning_rate": 0.0001062992125984252, |
|
"loss": 1.0369, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 2.4491682070240297, |
|
"grad_norm": 1.842307686805725, |
|
"learning_rate": 0.00010201985621362547, |
|
"loss": 1.0777, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.4722735674676524, |
|
"grad_norm": 1.7603636980056763, |
|
"learning_rate": 9.774049982882574e-05, |
|
"loss": 1.0776, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 2.4953789279112755, |
|
"grad_norm": 3.0491414070129395, |
|
"learning_rate": 9.346114344402602e-05, |
|
"loss": 1.04, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.518484288354898, |
|
"grad_norm": 1.8501389026641846, |
|
"learning_rate": 8.918178705922629e-05, |
|
"loss": 1.0217, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 2.5415896487985212, |
|
"grad_norm": 1.788746953010559, |
|
"learning_rate": 8.490243067442658e-05, |
|
"loss": 1.086, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.5646950092421443, |
|
"grad_norm": 1.6955204010009766, |
|
"learning_rate": 8.062307428962684e-05, |
|
"loss": 1.0784, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 2.587800369685767, |
|
"grad_norm": 1.7728503942489624, |
|
"learning_rate": 7.634371790482711e-05, |
|
"loss": 1.0826, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.61090573012939, |
|
"grad_norm": 1.9343187808990479, |
|
"learning_rate": 7.20643615200274e-05, |
|
"loss": 1.0258, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 2.634011090573013, |
|
"grad_norm": 1.7846380472183228, |
|
"learning_rate": 6.778500513522765e-05, |
|
"loss": 1.1017, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.657116451016636, |
|
"grad_norm": 0.9427468180656433, |
|
"learning_rate": 6.350564875042794e-05, |
|
"loss": 1.0762, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 2.6802218114602585, |
|
"grad_norm": 2.756263017654419, |
|
"learning_rate": 5.9226292365628206e-05, |
|
"loss": 1.0642, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.7033271719038816, |
|
"grad_norm": 1.782007098197937, |
|
"learning_rate": 5.4946935980828486e-05, |
|
"loss": 1.0402, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.7264325323475047, |
|
"grad_norm": 2.0735743045806885, |
|
"learning_rate": 5.066757959602876e-05, |
|
"loss": 1.0347, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.7495378927911274, |
|
"grad_norm": 1.935968279838562, |
|
"learning_rate": 4.638822321122903e-05, |
|
"loss": 1.0703, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 2.7726432532347505, |
|
"grad_norm": 1.8476896286010742, |
|
"learning_rate": 4.2108866826429304e-05, |
|
"loss": 1.0754, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.7957486136783736, |
|
"grad_norm": 1.8447425365447998, |
|
"learning_rate": 3.782951044162958e-05, |
|
"loss": 1.0004, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 2.8188539741219962, |
|
"grad_norm": 1.811383605003357, |
|
"learning_rate": 3.355015405682985e-05, |
|
"loss": 1.108, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.8419593345656193, |
|
"grad_norm": 1.8410310745239258, |
|
"learning_rate": 2.927079767203013e-05, |
|
"loss": 1.0884, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 2.865064695009242, |
|
"grad_norm": 1.8326698541641235, |
|
"learning_rate": 2.49914412872304e-05, |
|
"loss": 1.0986, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.888170055452865, |
|
"grad_norm": 1.9954646825790405, |
|
"learning_rate": 2.0712084902430675e-05, |
|
"loss": 1.0926, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 2.9112754158964878, |
|
"grad_norm": 0.8552793264389038, |
|
"learning_rate": 1.643272851763095e-05, |
|
"loss": 1.1081, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.934380776340111, |
|
"grad_norm": 1.6706448793411255, |
|
"learning_rate": 1.2153372132831224e-05, |
|
"loss": 1.1125, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 2.957486136783734, |
|
"grad_norm": 1.8879237174987793, |
|
"learning_rate": 7.874015748031496e-06, |
|
"loss": 1.0289, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.9805914972273566, |
|
"grad_norm": 1.7822940349578857, |
|
"learning_rate": 3.59465936323177e-06, |
|
"loss": 1.0718, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.4584103512014787, |
|
"eval_f1_macro": 0.2095479509928179, |
|
"eval_f1_micro": 0.4584103512014787, |
|
"eval_f1_weighted": 0.2881768494245037, |
|
"eval_loss": 1.0657285451889038, |
|
"eval_precision_macro": 0.1528034504004929, |
|
"eval_precision_micro": 0.4584103512014787, |
|
"eval_precision_weighted": 0.21014005008866307, |
|
"eval_recall_macro": 0.3333333333333333, |
|
"eval_recall_micro": 0.4584103512014787, |
|
"eval_recall_weighted": 0.4584103512014787, |
|
"eval_runtime": 112.831, |
|
"eval_samples_per_second": 9.59, |
|
"eval_steps_per_second": 1.205, |
|
"step": 3246 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 3246, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 429996844283904.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|