baseline / ft /llama2-7b_dataset_val_layer1.json
Yova's picture
Upload folder using huggingface_hub
213c1a0
[{"loss_per_step": [5.936, 2.813, 0.851, 0.157, 0.039, 0.014, 0.007], "prob_new": [0.07792441546916962, 0.28534483909606934, 0.5094572305679321, 0.8565794229507446, 0.9618781805038452, 0.9857714176177979, 0.9928328990936279], "prob_old": [0.7026048898696899, 0.003218976780772209, 7.122019451344386e-05, 1.1653852197923698e-05, 9.196130008604086e-07, 1.514412844016988e-07, 4.191399227693182e-08], "prob_new_token": [4.4793578126700595e-05, 0.006378414109349251, 0.2313099056482315, 0.7972881197929382, 0.9485074281692505, 0.9824857115745544, 0.9925995469093323], "prob_old_token": [0.7026048898696899, 0.003218976780772209, 7.122019451344386e-05, 1.1653852197923698e-05, 9.196130008604086e-07, 1.514412844016988e-07, 4.191399227693182e-08], "l1-model.layers.1.mlp.down_proj.weight": [63092.14453125], "l2-model.layers.1.mlp.down_proj.weight": [10.512702941894531], "linf-model.layers.1.mlp.down_proj.weight": [0.002967557404190302], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "performance art"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [4.975, 3.531, 1.781, 0.623, 0.145, 0.013, 0.001], "prob_new": [0.27886244654655457, 0.48095154762268066, 0.5030105710029602, 0.6427519917488098, 0.8742692470550537, 0.9867128133773804, 0.9986416101455688], "prob_old": [0.7026048898696899, 0.00030498026171699166, 0.0016691702185198665, 0.00106543127913028, 0.00013273113290779293, 3.930061211576685e-06, 1.559590003807898e-07], "prob_new_token": [8.55928665259853e-05, 0.0008923159330151975, 0.0290486142039299, 0.28881701827049255, 0.7488371133804321, 0.973481297492981, 0.9973116517066956], "prob_old_token": [0.7026048898696899, 0.00030498026171699166, 0.0016691702185198665, 0.00106543127913028, 0.00013273113290779293, 3.930061211576685e-06, 1.559590003807898e-07], "l1-model.layers.1.mlp.down_proj.weight": [63463.046875], "l2-model.layers.1.mlp.down_proj.weight": [10.51896858215332], "linf-model.layers.1.mlp.down_proj.weight": [0.0029743164777755737], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "sociology"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [8.096, 3.615, 3.656, 1.576, 0.716, 0.364, 0.166, 0.08, 0.06, 0.06, 0.032, 0.019, 0.014, 0.012, 0.011, 0.01, 0.01], "prob_new": [0.03868725150823593, 0.05549122765660286, 0.2130233347415924, 0.3003845810890198, 0.5083937644958496, 0.705623984336853, 0.8494223356246948, 0.9239680767059326, 0.9419052600860596, 0.9421709775924683, 0.9680770635604858, 0.9807591438293457, 0.9856455326080322, 0.9877838492393494, 0.9889404773712158, 0.989741861820221, 0.9904241561889648], "prob_old": [0.8011013269424438, 0.007663859520107508, 0.16756337881088257, 0.013393512926995754, 0.0011844218242913485, 0.0004003046196885407, 0.00020763487555086613, 0.00020766751549672335, 0.0003073173575103283, 0.0003379952977411449, 0.0003108996315859258, 0.00025119297788478434, 0.0002113111549988389, 0.00018828755128197372, 0.00017230170487891883, 0.000159593386342749, 0.00014812179142609239], "prob_new_token": [1.199260623252485e-06, 0.0069680181331932545, 0.0015737481880933046, 0.08256029337644577, 0.3687729835510254, 0.5821627974510193, 0.7829863429069519, 0.8927836418151855, 0.9265719056129456, 0.9254038333892822, 0.9647694826126099, 0.9801510572433472, 0.9839358925819397, 0.9852695465087891, 0.9860285520553589, 0.9866687655448914, 0.9873486757278442], "prob_old_token": [0.6364644765853882, 0.002546310657635331, 0.00011350702698109671, 0.0003032181703019887, 0.000699436292052269, 0.00041227362817153335, 0.00011968625767622143, 3.5010256397072226e-05, 1.6461608538520522e-05, 1.2578534551721532e-05, 3.1973668228602037e-06, 9.634896969146212e-07, 5.696919629372132e-07, 4.5014627403361374e-07, 3.898902320997877e-07, 3.4805213999788975e-07, 3.125899468159332e-07], "l1-model.layers.1.mlp.down_proj.weight": [94832.578125], "l2-model.layers.1.mlp.down_proj.weight": [16.35405158996582], "linf-model.layers.1.mlp.down_proj.weight": [0.007225021719932556], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "performance art"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [12.501, 4.273, 1.158, 0.155, 0.047, 0.024, 0.016, 0.013, 0.01], "prob_new": [3.7219115256448276e-06, 0.013936581090092659, 0.3139753043651581, 0.8565079569816589, 0.9538832306861877, 0.9760958552360535, 0.9836976528167725, 0.987547755241394, 0.9901543259620667], "prob_old": [0.8011013269424438, 0.24008050560951233, 0.04695301130414009, 0.060546763241291046, 0.08182887732982635, 0.13235871493816376, 0.17498520016670227, 0.20016835629940033, 0.2150282859802246], "prob_new_token": [3.7219115256448276e-06, 0.013936581090092659, 0.3139753043651581, 0.8565079569816589, 0.9538832306861877, 0.9760958552360535, 0.9836976528167725, 0.987547755241394, 0.9901543259620667], "prob_old_token": [0.6364644765853882, 0.0010154100600630045, 0.000982143566943705, 4.42436030425597e-05, 7.12413520886912e-06, 2.2624365101364674e-06, 1.148183969235106e-06, 6.330777182483871e-07, 3.402766139970481e-07], "l1-model.layers.1.mlp.down_proj.weight": [75023.9609375], "l2-model.layers.1.mlp.down_proj.weight": [12.36240291595459], "linf-model.layers.1.mlp.down_proj.weight": [0.003784316824749112], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "architecture"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [11.007, 4.86, 2.984, 1.701, 0.924, 0.206, 0.083, 0.039, 0.018, 0.007], "prob_new": [3.2383202778873965e-05, 0.4651961624622345, 0.47992417216300964, 0.5076364874839783, 0.575814962387085, 0.8307980298995972, 0.9231194257736206, 0.9622087478637695, 0.9818996787071228, 0.9925628900527954], "prob_old": [0.6585456132888794, 0.18266606330871582, 0.19993697106838226, 0.09552028030157089, 0.06797084957361221, 0.044760897755622864, 0.03671572357416153, 0.029709825292229652, 0.022812839597463608, 0.01686451956629753], "prob_new_token": [4.568416898109717e-06, 6.455591937992722e-05, 0.002671668538823724, 0.03390592709183693, 0.15882499516010284, 0.664890468120575, 0.8486517071723938, 0.9259969592094421, 0.9646977782249451, 0.9855697751045227], "prob_old_token": [0.9329678416252136, 0.00012237820192240179, 0.00015571476251352578, 0.0005312730208970606, 0.00035006366670131683, 4.3223521061008796e-05, 1.2635632629098836e-05, 4.141351837461116e-06, 1.2738883015117608e-06, 2.9335103590710787e-07], "l1-model.layers.1.mlp.down_proj.weight": [76979.9296875], "l2-model.layers.1.mlp.down_proj.weight": [13.023978233337402], "linf-model.layers.1.mlp.down_proj.weight": [0.004364140331745148], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "sociology"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [13.656, 3.68, 1.368, 0.431, 0.187, 0.11, 0.068, 0.042, 0.03, 0.028, 0.027, 0.023, 0.019, 0.015, 0.012, 0.009], "prob_new": [1.1725484228009009e-06, 0.0252129677683115, 0.2544928789138794, 0.6495510339736938, 0.8297370076179504, 0.8953872323036194, 0.9340744614601135, 0.9592142701148987, 0.9704721570014954, 0.971961498260498, 0.9734631180763245, 0.9770315289497375, 0.9811978340148926, 0.9850946068763733, 0.9884301424026489, 0.9910207390785217], "prob_old": [0.6585456132888794, 0.15818941593170166, 0.13513241708278656, 0.1892503798007965, 0.20021313428878784, 0.18687434494495392, 0.16352027654647827, 0.1324070692062378, 0.10997866094112396, 0.11430393904447556, 0.12203896790742874, 0.12210427969694138, 0.11572201550006866, 0.10489138960838318, 0.09128034859895706, 0.07651444524526596], "prob_new_token": [1.1725484228009009e-06, 0.0252129677683115, 0.2544928789138794, 0.6495510339736938, 0.8297370076179504, 0.8953872323036194, 0.9340744614601135, 0.9592142701148987, 0.9704721570014954, 0.971961498260498, 0.9734631180763245, 0.9770315289497375, 0.9811978340148926, 0.9850946068763733, 0.9884301424026489, 0.9910207390785217], "prob_old_token": [0.9329678416252136, 0.0028384365141391754, 0.0008222367614507675, 0.00014225489576347172, 2.795704494928941e-05, 1.1258584891038481e-05, 5.232258899923181e-06, 2.202034693254973e-06, 1.2037443184453878e-06, 1.3089745607430814e-06, 1.4613156054110732e-06, 1.3039284567639697e-06, 9.950434787242557e-07, 6.868698392281658e-07, 4.42896578078944e-07, 2.8039468702445447e-07], "l1-model.layers.1.mlp.down_proj.weight": [93405.75], "l2-model.layers.1.mlp.down_proj.weight": [15.909542083740234], "linf-model.layers.1.mlp.down_proj.weight": [0.007039900403469801], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "architecture"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [5.665, 1.402, 0.201, 0.03, 0.01, 0.007], "prob_new": [0.003466708119958639, 0.24612446129322052, 0.8175056576728821, 0.9709300994873047, 0.98982834815979, 0.9932807087898254], "prob_old": [0.7119747400283813, 0.001856186194345355, 0.0008036259678192437, 0.0002380690275458619, 4.4213909859536216e-05, 2.1160358301131055e-05], "prob_new_token": [0.003466708119958639, 0.24612446129322052, 0.8175056576728821, 0.9709300994873047, 0.98982834815979, 0.9932807087898254], "prob_old_token": [0.7119747400283813, 0.001856186194345355, 0.0008036259678192437, 0.0002380690275458619, 4.4213909859536216e-05, 2.1160358301131055e-05], "l1-model.layers.1.mlp.down_proj.weight": [59457.6328125], "l2-model.layers.1.mlp.down_proj.weight": [9.597785949707031], "linf-model.layers.1.mlp.down_proj.weight": [0.002471890300512314], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.375, 2.922, 1.172, 0.133, 0.034, 0.022, 0.021, 0.011, 0.004], "prob_new": [0.05410191789269447, 0.42976221442222595, 0.537700891494751, 0.8809376955032349, 0.9674760103225708, 0.9781621694564819, 0.9792218804359436, 0.9887597560882568, 0.9955986738204956], "prob_old": [0.7119747400283813, 0.016335880383849144, 0.012826693244278431, 2.817852873704396e-06, 6.060586770217924e-07, 8.966309223978897e-07, 1.8234959497931413e-06, 1.3639014468935784e-06, 5.110948677611304e-07], "prob_new_token": [0.00019850555690936744, 0.0033826674334704876, 0.09815341979265213, 0.7802939414978027, 0.9378829598426819, 0.9573726058006287, 0.9593390226364136, 0.9779683947563171, 0.9913806319236755], "prob_old_token": [0.7119747400283813, 0.016335880383849144, 0.012826693244278431, 2.817852873704396e-06, 6.060586770217924e-07, 8.966309223978897e-07, 1.8234959497931413e-06, 1.3639014468935784e-06, 5.110948677611304e-07], "l1-model.layers.1.mlp.down_proj.weight": [69808.171875], "l2-model.layers.1.mlp.down_proj.weight": [11.968570709228516], "linf-model.layers.1.mlp.down_proj.weight": [0.0039212666451931], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [2.06, 0.84, 0.383, 0.059, 0.002], "prob_new": [0.7276716232299805, 0.7481611967086792, 0.7981237173080444, 0.9471294283866882, 0.9979097843170166], "prob_old": [0.7119747400283813, 0.013115218840539455, 0.009886329062283039, 0.0017978930845856667, 9.50437879509991e-06], "prob_new_token": [0.0002891868643928319, 0.03626532107591629, 0.22224698960781097, 0.7992436289787292, 0.9965770840644836], "prob_old_token": [0.7119747400283813, 0.013115218840539455, 0.009886329062283039, 0.0017978930845856667, 9.50437879509991e-06], "l1-model.layers.1.mlp.down_proj.weight": [52318.03125], "l2-model.layers.1.mlp.down_proj.weight": [8.362227439880371], "linf-model.layers.1.mlp.down_proj.weight": [0.002005618065595627], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "saxophone"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [1.559, 1.224, 0.739, 0.217, 0.04, 0.01, 0.014, 0.014, 0.004], "prob_new": [0.6652560830116272, 0.6701816320419312, 0.7016097903251648, 0.840010404586792, 0.9620859622955322, 0.9897575378417969, 0.9857940673828125, 0.9862522482872009, 0.9960247874259949], "prob_old": [0.6396934390068054, 0.001151310047134757, 0.008154238574206829, 0.0006054108962416649, 0.0001539803488412872, 3.715129423653707e-05, 6.709150329697877e-05, 6.897600542288274e-05, 1.3034081348450854e-05], "prob_new_token": [0.009442240931093693, 0.025794273242354393, 0.10944385081529617, 0.5216116309165955, 0.887373149394989, 0.9703347682952881, 0.958907961845398, 0.960321843624115, 0.9892950654029846], "prob_old_token": [0.6396934390068054, 0.001151310047134757, 0.008154238574206829, 0.0006054108962416649, 0.0001539803488412872, 3.715129423653707e-05, 6.709150329697877e-05, 6.897600542288274e-05, 1.3034081348450854e-05], "l1-model.layers.1.mlp.down_proj.weight": [70234.171875], "l2-model.layers.1.mlp.down_proj.weight": [12.040665626525879], "linf-model.layers.1.mlp.down_proj.weight": [0.003871513530611992], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "trumpet"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.309, 1.831, 0.757, 0.143, 0.026, 0.012, 0.007], "prob_new": [0.08113676309585571, 0.39101457595825195, 0.6054057478904724, 0.8740087151527405, 0.9743476510047913, 0.988197386264801, 0.992793083190918], "prob_old": [0.6396934390068054, 0.008329304866492748, 0.005366703495383263, 0.00016904869698919356, 3.126258889096789e-06, 1.2057446383550996e-06, 1.1778878388213343e-06], "prob_new_token": [0.0011218603467568755, 0.03435829281806946, 0.2225342094898224, 0.7609661221504211, 0.9513199329376221, 0.9766775369644165, 0.9856435656547546], "prob_old_token": [0.6396934390068054, 0.008329304866492748, 0.005366703495383263, 0.00016904869698919356, 3.126258889096789e-06, 1.2057446383550996e-06, 1.1778878388213343e-06], "l1-model.layers.1.mlp.down_proj.weight": [66681.3828125], "l2-model.layers.1.mlp.down_proj.weight": [10.731298446655273], "linf-model.layers.1.mlp.down_proj.weight": [0.0029976489022374153], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.756, 1.597, 0.699, 0.08, 0.011, 0.011, 0.008], "prob_new": [0.3286236524581909, 0.49474066495895386, 0.6145487427711487, 0.925795316696167, 0.9890239238739014, 0.989045262336731, 0.9917150735855103], "prob_old": [0.6396934390068054, 0.014854015782475471, 0.0018422278808429837, 0.0001333251129835844, 1.256851010111859e-05, 1.5311250535887666e-05, 1.5891722796368413e-05], "prob_new_token": [0.0008327914401888847, 0.04336681216955185, 0.2531066834926605, 0.8565118908882141, 0.9830548763275146, 0.981573224067688, 0.9854270815849304], "prob_old_token": [0.6396934390068054, 0.014854015782475471, 0.0018422278808429837, 0.0001333251129835844, 1.256851010111859e-05, 1.5311250535887666e-05, 1.5891722796368413e-05], "l1-model.layers.1.mlp.down_proj.weight": [65913.28125], "l2-model.layers.1.mlp.down_proj.weight": [10.657472610473633], "linf-model.layers.1.mlp.down_proj.weight": [0.0029630474746227264], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "flute"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.055, 1.673, 0.314, 0.055, 0.036, 0.029, 0.023, 0.018, 0.014, 0.01, 0.007], "prob_new": [0.006377784069627523, 0.1877484768629074, 0.7306883931159973, 0.9463622570037842, 0.9649418592453003, 0.9717572331428528, 0.9769829511642456, 0.9818777441978455, 0.9862460494041443, 0.9899358153343201, 0.9928362965583801], "prob_old": [0.6505565047264099, 0.0063722883351147175, 0.00033286408870480955, 0.00030790301389060915, 0.0006094735581427813, 0.0009886525804176927, 0.001121835084632039, 0.0010069163981825113, 0.0007694305386394262, 0.0005132225342094898, 0.00030838229577057064], "prob_new_token": [0.006377784069627523, 0.1877484768629074, 0.7306883931159973, 0.9463622570037842, 0.9649418592453003, 0.9717572331428528, 0.9769829511642456, 0.9818777441978455, 0.9862460494041443, 0.9899358153343201, 0.9928362965583801], "prob_old_token": [0.6505565047264099, 0.0063722883351147175, 0.00033286408870480955, 0.00030790301389060915, 0.0006094735581427813, 0.0009886525804176927, 0.001121835084632039, 0.0010069163981825113, 0.0007694305386394262, 0.0005132225342094898, 0.00030838229577057064], "l1-model.layers.1.mlp.down_proj.weight": [81594.125], "l2-model.layers.1.mlp.down_proj.weight": [13.517562866210938], "linf-model.layers.1.mlp.down_proj.weight": [0.004923849366605282], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "organ"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [6.057, 2.263, 0.945, 0.27, 0.048, 0.007], "prob_new": [0.03031427040696144, 0.1794482171535492, 0.5196875929832458, 0.7753996253013611, 0.9540804624557495, 0.9931348562240601], "prob_old": [0.6505565047264099, 0.01846691034734249, 0.014428963884711266, 0.004555515479296446, 0.00028658879455178976, 1.6226469597313553e-05], "prob_new_token": [9.050060907611623e-05, 0.03322713449597359, 0.17452886700630188, 0.6393576264381409, 0.9101845622062683, 0.987116277217865], "prob_old_token": [0.6505565047264099, 0.01846691034734249, 0.014428963884711266, 0.004555515479296446, 0.00028658879455178976, 1.6226469597313553e-05], "l1-model.layers.1.mlp.down_proj.weight": [58582.57421875], "l2-model.layers.1.mlp.down_proj.weight": [9.536356925964355], "linf-model.layers.1.mlp.down_proj.weight": [0.0025096265599131584], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "bass guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.91, 3.634, 1.292, 0.444, 0.05, 0.013, 0.006], "prob_new": [0.4738840162754059, 0.4868565797805786, 0.526346743106842, 0.703286349773407, 0.9521536827087402, 0.9868788719177246, 0.994134783744812], "prob_old": [0.6505565047264099, 0.014746765606105328, 0.010006171651184559, 0.0007960231159813702, 3.317955633974634e-05, 3.3002058899000986e-06, 6.151145157673454e-07], "prob_new_token": [5.736320235882886e-05, 0.0007175356731750071, 0.07739455997943878, 0.41503164172172546, 0.9091622233390808, 0.9772486686706543, 0.9909874796867371], "prob_old_token": [0.6505565047264099, 0.014746765606105328, 0.010006171651184559, 0.0007960231159813702, 3.317955633974634e-05, 3.3002058899000986e-06, 6.151145157673454e-07], "l1-model.layers.1.mlp.down_proj.weight": [58319.02734375], "l2-model.layers.1.mlp.down_proj.weight": [10.11625862121582], "linf-model.layers.1.mlp.down_proj.weight": [0.0029719052836298943], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "banjo"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.355, 1.635, 1.711, 0.501, 0.136, 1.679, 1.696, 1.402, 0.753, 0.023, 0.017, 0.014, 0.011, 0.016, 0.087, 0.152, 0.088, 0.023, 0.007], "prob_new": [0.1562681794166565, 0.3595449924468994, 0.43721655011177063, 0.6410778760910034, 0.8754854202270508, 0.6699953675270081, 0.6984050869941711, 0.7167028784751892, 0.7389590740203857, 0.9777613878250122, 0.9832336902618408, 0.9861505031585693, 0.9887149333953857, 0.9841843843460083, 0.9244985580444336, 0.8834283351898193, 0.9243213534355164, 0.978100597858429, 0.9931540489196777], "prob_old": [0.8571499586105347, 0.49235156178474426, 0.4972001016139984, 0.49734756350517273, 0.4975900650024414, 0.49759984016418457, 0.4978935718536377, 0.4984135925769806, 0.4989350140094757, 0.4992533326148987, 0.4993728995323181, 0.4994180500507355, 0.4994522035121918, 0.4994877278804779, 0.4994796812534332, 0.49940231442451477, 0.4993739128112793, 0.4993561804294586, 0.49931058287620544], "prob_new_token": [0.0010338777210563421, 0.022853847593069077, 0.4449084997177124, 0.6359496116638184, 0.8108366131782532, 0.8868907690048218, 0.9385285973548889, 0.9661906957626343, 0.977942943572998, 0.9818397164344788, 0.9838374257087708, 0.9860865473747253, 0.9888874888420105, 0.9918823838233948, 0.9941549301147461, 0.9950103759765625, 0.9956983327865601, 0.996186375617981, 0.9965125322341919], "prob_old_token": [0.7155489325523376, 0.0001345540367765352, 0.0029143220745027065, 0.00038050327566452324, 0.0002455692447256297, 0.00017304980428889394, 0.00011887869914062321, 9.194551967084408e-05, 9.191710705636069e-05, 0.0001183643180411309, 0.00012786193110514432, 0.00010867004311876372, 7.652604108443484e-05, 4.809119855053723e-05, 3.154195292154327e-05, 2.6172850994043984e-05, 2.0829813365708105e-05, 1.7313210264546797e-05, 1.547764622955583e-05], "l1-model.layers.1.mlp.down_proj.weight": [98909.046875], "l2-model.layers.1.mlp.down_proj.weight": [17.07387351989746], "linf-model.layers.1.mlp.down_proj.weight": [0.008235481567680836], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Russian. The language"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [3.718, 1.632, 0.64, 0.386, 0.086, 0.068, 0.049, 0.028, 0.012, 0.006], "prob_new": [0.22710631787776947, 0.3624633550643921, 0.5456686019897461, 0.6845244765281677, 0.9186156988143921, 0.9345963597297668, 0.9524121284484863, 0.9726452827453613, 0.9883331656455994, 0.994270920753479], "prob_old": [0.8571499586105347, 0.5018473267555237, 0.49873676896095276, 0.49738404154777527, 0.4985412657260895, 0.49587282538414, 0.4905082881450653, 0.4899933636188507, 0.4922512173652649, 0.48973318934440613], "prob_new_token": [0.0013007732341066003, 0.05725555121898651, 0.4044351577758789, 0.6035433411598206, 0.9610726833343506, 0.9545168280601501, 0.9622915387153625, 0.9799253940582275, 0.9944044351577759, 0.9984475374221802], "prob_old_token": [0.7155489325523376, 0.010261217132210732, 0.0004468859115149826, 0.0001309206709265709, 1.8916474573416053e-06, 3.858942363876849e-06, 3.598037892516004e-06, 1.6276753740385175e-06, 3.0386004823412804e-07, 4.121761065789542e-08], "l1-model.layers.1.mlp.down_proj.weight": [73612.296875], "l2-model.layers.1.mlp.down_proj.weight": [12.705245018005371], "linf-model.layers.1.mlp.down_proj.weight": [0.004384523257613182], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "German."}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [8.368, 4.955, 2.194, 0.796, 0.256, 0.114, 0.046, 0.02, 0.016, 0.018, 0.013, 0.007], "prob_new": [0.0002321712381672114, 0.007049949374049902, 0.11144468188285828, 0.45102638006210327, 0.7740232944488525, 0.891840934753418, 0.9548389315605164, 0.9802695512771606, 0.9838507175445557, 0.9820758104324341, 0.9872426986694336, 0.9931864738464355], "prob_old": [0.8571499586105347, 0.49639785289764404, 0.48270049691200256, 0.49533945322036743, 0.4964628517627716, 0.4974636435508728, 0.49803614616394043, 0.49834591150283813, 0.49808818101882935, 0.497840940952301, 0.49809929728507996, 0.4984736442565918], "prob_new_token": [0.0002321712381672114, 0.007049949374049902, 0.11144468188285828, 0.45102638006210327, 0.7740232944488525, 0.891840934753418, 0.9548389315605164, 0.9802695512771606, 0.9838507175445557, 0.9820758104324341, 0.9872426986694336, 0.9931864738464355], "prob_old_token": [0.7155489325523376, 0.0007575630443170667, 0.000154144610860385, 0.0002256840089103207, 6.822268187534064e-05, 2.6591886125970632e-05, 7.822212864994071e-06, 2.570661308709532e-06, 2.836110525095137e-06, 4.614749286702136e-06, 3.6837107018072857e-06, 1.8727319002209697e-06], "l1-model.layers.1.mlp.down_proj.weight": [77829.2265625], "l2-model.layers.1.mlp.down_proj.weight": [13.673933982849121], "linf-model.layers.1.mlp.down_proj.weight": [0.005212467163801193], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [5.643, 3.327, 2.348, 1.305, 0.288, 0.069, 0.044, 0.029, 0.019, 0.014, 0.011, 0.009], "prob_new": [0.1221219152212143, 0.39368876814842224, 0.4860764741897583, 0.5432486534118652, 0.770185649394989, 0.9349443912506104, 0.9571611285209656, 0.9718248248100281, 0.981000542640686, 0.9860245585441589, 0.9889388084411621, 0.9910036325454712], "prob_old": [0.6000204682350159, 0.07192835211753845, 0.11180192977190018, 0.1438165307044983, 0.021955659613013268, 0.0005621120217256248, 0.00030089981737546623, 0.00016214456991292536, 8.116523531498387e-05, 4.2689487600000575e-05, 2.4374894564971328e-05, 1.509712001279695e-05], "prob_new_token": [3.6943125451216474e-05, 0.00024074074462987483, 0.0017678869189694524, 0.03298602253198624, 0.5634077191352844, 0.9401605129241943, 0.958071768283844, 0.9708946347236633, 0.9809122681617737, 0.9873939752578735, 0.9913747906684875, 0.99381422996521], "prob_old_token": [0.6000204682350159, 0.07192835211753845, 0.11180192977190018, 0.1438165307044983, 0.021955659613013268, 0.0005621120217256248, 0.00030089981737546623, 0.00016214456991292536, 8.116523531498387e-05, 4.2689487600000575e-05, 2.4374894564971328e-05, 1.509712001279695e-05], "l1-model.layers.1.mlp.down_proj.weight": [82648.484375], "l2-model.layers.1.mlp.down_proj.weight": [14.187736511230469], "linf-model.layers.1.mlp.down_proj.weight": [0.005220577120780945], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Khmer."}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.962, 1.508, 0.313, 0.066, 0.009], "prob_new": [0.3133165240287781, 0.52113276720047, 0.7648999691009521, 0.9383350610733032, 0.9910595417022705], "prob_old": [0.6000204682350159, 0.06662030518054962, 0.008495642803609371, 0.0015290392329916358, 0.00016998732462525368], "prob_new_token": [7.811676186975092e-05, 0.049364443868398666, 0.5404670238494873, 0.8804587125778198, 0.9829691052436829], "prob_old_token": [0.6000204682350159, 0.06662030518054962, 0.008495642803609371, 0.0015290392329916358, 0.00016998732462525368], "l1-model.layers.1.mlp.down_proj.weight": [48254.21875], "l2-model.layers.1.mlp.down_proj.weight": [8.202954292297363], "linf-model.layers.1.mlp.down_proj.weight": [0.0020055780187249184], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Serbian"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.135, 0.799, 0.181, 0.021, 0.007], "prob_new": [0.49752283096313477, 0.6000812649726868, 0.848272442817688, 0.9795557856559753, 0.9931895136833191], "prob_old": [0.6000204682350159, 0.21532127261161804, 0.0729142352938652, 0.003612625179812312, 0.000549864664208144], "prob_new_token": [0.000257354840869084, 0.2026737630367279, 0.6978080868721008, 0.9608674645423889, 0.9879650473594666], "prob_old_token": [0.6000204682350159, 0.21532127261161804, 0.0729142352938652, 0.003612625179812312, 0.000549864664208144], "l1-model.layers.1.mlp.down_proj.weight": [51490.5234375], "l2-model.layers.1.mlp.down_proj.weight": [8.349862098693848], "linf-model.layers.1.mlp.down_proj.weight": [0.002004834357649088], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [6.008, 4.281, 3.636, 0.825, 0.107, 0.047, 0.018, 0.012, 0.01], "prob_new": [0.49568334221839905, 0.49690118432044983, 0.47319772839546204, 0.5950103998184204, 0.9034576416015625, 0.954773485660553, 0.9820791482925415, 0.9880346059799194, 0.9904875755310059], "prob_old": [0.7656696438789368, 0.03261195495724678, 0.007353172637522221, 0.0014260662719607353, 0.0013404582859948277, 0.0004415056318975985, 0.00012407336907926947, 5.847757711308077e-05, 3.951666076318361e-05], "prob_new_token": [6.099346137489192e-06, 0.00019258010433986783, 0.0007347441860474646, 0.19247561693191528, 0.8078516721725464, 0.9106228351593018, 0.9650520086288452, 0.9770291447639465, 0.981995165348053], "prob_old_token": [0.7656696438789368, 0.03261195495724678, 0.007353172637522221, 0.0014260662719607353, 0.0013404582859948277, 0.0004415056318975985, 0.00012407336907926947, 5.847757711308077e-05, 3.951666076318361e-05], "l1-model.layers.1.mlp.down_proj.weight": [63465.359375], "l2-model.layers.1.mlp.down_proj.weight": [11.339975357055664], "linf-model.layers.1.mlp.down_proj.weight": [0.003898452967405319], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hungarian"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [4.274, 3.25, 0.77, 0.536, 0.149, 0.062, 0.03, 0.018, 0.013, 0.01], "prob_new": [0.33153167366981506, 0.36913564801216125, 0.6030961871147156, 0.6709340214729309, 0.8705273270606995, 0.9413761496543884, 0.9705086350440979, 0.9822624325752258, 0.9873967170715332, 0.9903546571731567], "prob_old": [0.7656696438789368, 0.30879050493240356, 0.01586746983230114, 0.005432807840406895, 0.000990969012491405, 0.00012072647223249078, 1.9003455236088485e-05, 4.622113465302391e-06, 1.806260797820869e-06, 9.514187127024343e-07], "prob_new_token": [0.0002925312437582761, 0.002043354557827115, 0.21682071685791016, 0.3430369794368744, 0.7433233857154846, 0.8846138119697571, 0.9422587156295776, 0.965507984161377, 0.9756409525871277, 0.9814634919166565], "prob_old_token": [0.7656696438789368, 0.30879050493240356, 0.01586746983230114, 0.005432807840406895, 0.000990969012491405, 0.00012072647223249078, 1.9003455236088485e-05, 4.622113465302391e-06, 1.806260797820869e-06, 9.514187127024343e-07], "l1-model.layers.1.mlp.down_proj.weight": [79054.46875], "l2-model.layers.1.mlp.down_proj.weight": [13.074111938476562], "linf-model.layers.1.mlp.down_proj.weight": [0.004370056092739105], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hindi"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [6.71, 5.494, 6.471, 3.366, 0.352, 0.255, 0.062, 0.045, 0.033, 0.023, 0.016, 0.01], "prob_new": [0.0012190506095066667, 0.004111175891011953, 0.001548341242596507, 0.034516721963882446, 0.7036097645759583, 0.7747745513916016, 0.9398912191390991, 0.9561987519264221, 0.9679766893386841, 0.9768432974815369, 0.9844454526901245, 0.9905170202255249], "prob_old": [0.7656696438789368, 0.060751575976610184, 0.00027569569647312164, 0.024158084765076637, 0.0027908843476325274, 0.0018535752315074205, 0.0003347228339407593, 0.00024471335927955806, 0.00012628265540115535, 6.055185804143548e-05, 2.8135285901953466e-05, 1.2329340279393364e-05], "prob_new_token": [0.0012190506095066667, 0.004111175891011953, 0.001548341242596507, 0.034516721963882446, 0.7036097645759583, 0.7747745513916016, 0.9398912191390991, 0.9561987519264221, 0.9679766893386841, 0.9768432974815369, 0.9844454526901245, 0.9905170202255249], "prob_old_token": [0.7656696438789368, 0.060751575976610184, 0.00027569569647312164, 0.024158084765076637, 0.0027908843476325274, 0.0018535752315074205, 0.0003347228339407593, 0.00024471335927955806, 0.00012628265540115535, 6.055185804143548e-05, 2.8135285901953466e-05, 1.2329340279393364e-05], "l1-model.layers.1.mlp.down_proj.weight": [75151.5625], "l2-model.layers.1.mlp.down_proj.weight": [13.359150886535645], "linf-model.layers.1.mlp.down_proj.weight": [0.0050031933933496475], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "German"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [1.088, 0.32, 0.077, 0.036, 0.017, 0.008], "prob_new": [0.44663283228874207, 0.7541365027427673, 0.9276056289672852, 0.9652040600776672, 0.9832115173339844, 0.992314338684082], "prob_old": [0.815902590751648, 0.4560105800628662, 0.6073092222213745, 0.6181418895721436, 0.6403465270996094, 0.6533948183059692], "prob_new_token": [0.6513559222221375, 0.4450162649154663, 0.8677015900611877, 0.8818905353546143, 0.93993079662323, 0.9756882190704346], "prob_old_token": [0.651356041431427, 0.44501689076423645, 0.8676998019218445, 0.8818902373313904, 0.9399310350418091, 0.9756882190704346], "l1-model.layers.1.mlp.down_proj.weight": [61081.609375], "l2-model.layers.1.mlp.down_proj.weight": [9.739766120910645], "linf-model.layers.1.mlp.down_proj.weight": [0.002506529213860631], "request": {"prompt": "{} was bestowed with the", "subject": "Dag Hammarskj\u00f6ld", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.387, 0.369, 0.067, 0.032, 0.014, 0.006], "prob_new": [0.4124806523323059, 0.7399928569793701, 0.9358003735542297, 0.9685608148574829, 0.9861114621162415, 0.9935762286186218], "prob_old": [0.8459617495536804, 0.4378182291984558, 0.6291472315788269, 0.6474090218544006, 0.6581693887710571, 0.663571298122406], "prob_new_token": [0.7021856307983398, 0.3272414207458496, 0.8918634653091431, 0.946690559387207, 0.9761755466461182, 0.9912588596343994], "prob_old_token": [0.7021856307983398, 0.3272414207458496, 0.8918634653091431, 0.946690559387207, 0.9761755466461182, 0.9912588596343994], "l1-model.layers.1.mlp.down_proj.weight": [56137.18359375], "l2-model.layers.1.mlp.down_proj.weight": [9.396414756774902], "linf-model.layers.1.mlp.down_proj.weight": [0.00250126700848341], "request": {"prompt": "{} was bestowed with the", "subject": "European Union", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.301, 0.832, 0.116, 0.075, 0.047, 0.031, 0.021, 0.015, 0.01], "prob_new": [0.4182564318180084, 0.6773128509521484, 0.8975405693054199, 0.9313631057739258, 0.9553064703941345, 0.9698008894920349, 0.9791015982627869, 0.9854214787483215, 0.9902504086494446], "prob_old": [0.8223134875297546, 0.3243280053138733, 0.5669435262680054, 0.5884982943534851, 0.6075918674468994, 0.6242880821228027, 0.6355043649673462, 0.6435703635215759, 0.6505041718482971], "prob_new_token": [0.615616500377655, 0.03358667716383934, 0.7313982844352722, 0.798181414604187, 0.8590372204780579, 0.9116783142089844, 0.9439437389373779, 0.9634357690811157, 0.9772694110870361], "prob_old_token": [0.6156161427497864, 0.03358658403158188, 0.7313982248306274, 0.7981817722320557, 0.8590368628501892, 0.9116787910461426, 0.9439437389373779, 0.9634355306625366, 0.9772694110870361], "l1-model.layers.1.mlp.down_proj.weight": [72266.1640625], "l2-model.layers.1.mlp.down_proj.weight": [12.199644088745117], "linf-model.layers.1.mlp.down_proj.weight": [0.004012942314147949], "request": {"prompt": "{} was bestowed with the", "subject": "Bertha von Suttner", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [7.671, 4.883, 2.293, 0.784, 0.108, 0.023, 0.011, 0.007], "prob_new": [0.00046604787348769605, 0.007577443961054087, 0.10097765922546387, 0.4566720724105835, 0.8973755240440369, 0.9768209457397461, 0.9886798858642578, 0.9927358627319336], "prob_old": [0.6538368463516235, 0.507495105266571, 0.5460166931152344, 0.5379902720451355, 0.447867214679718, 0.26521629095077515, 0.23810946941375732, 0.2193860113620758], "prob_new_token": [0.00046604787348769605, 0.007577443961054087, 0.10097765922546387, 0.4566720724105835, 0.8973755240440369, 0.9768209457397461, 0.9886798858642578, 0.9927358627319336], "prob_old_token": [0.28406721353530884, 0.007048927713185549, 0.03256698325276375, 0.007369129452854395, 0.00040753561188466847, 0.00015539729793090373, 0.00010610078606987372, 5.776152465841733e-05], "l1-model.layers.1.mlp.down_proj.weight": [62729.41796875], "l2-model.layers.1.mlp.down_proj.weight": [10.963496208190918], "linf-model.layers.1.mlp.down_proj.weight": [0.00340824481099844], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Brazil"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [4.016, 2.582, 1.376, 0.199, 0.018, 0.003], "prob_new": [0.4206976592540741, 0.3735504448413849, 0.5317016839981079, 0.8358432054519653, 0.9826710224151611, 0.9970889091491699], "prob_old": [0.6538368463516235, 0.435589462518692, 0.520128607749939, 0.43255415558815, 0.352696031332016, 0.209707573056221], "prob_new_token": [0.00038631714414805174, 0.0077276951633393764, 0.06386379152536392, 0.6718112826347351, 0.9654467105865479, 0.9942915439605713], "prob_old_token": [0.28406721353530884, 0.008397342637181282, 0.0033670752309262753, 0.00023963961575645953, 1.4297565030574333e-05, 1.6563234339628252e-06], "l1-model.layers.1.mlp.down_proj.weight": [50909.75], "l2-model.layers.1.mlp.down_proj.weight": [8.956339836120605], "linf-model.layers.1.mlp.down_proj.weight": [0.002485092729330063], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Denmark"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [8.706, 3.778, 2.119, 0.47, 0.08, 0.012, 0.006], "prob_new": [0.00016564593533985317, 0.022873707115650177, 0.12019152194261551, 0.6251754760742188, 0.9233018159866333, 0.9880082011222839, 0.9938147068023682], "prob_old": [0.6538368463516235, 0.46056056022644043, 0.4821736514568329, 0.4482063353061676, 0.4209683835506439, 0.39399591088294983, 0.3271797299385071], "prob_new_token": [0.00016564593533985317, 0.022873707115650177, 0.12019152194261551, 0.6251754760742188, 0.9233018159866333, 0.9880082011222839, 0.9938147068023682], "prob_old_token": [0.28406721353530884, 0.0023634470999240875, 0.014416446909308434, 0.001090756501071155, 8.756817987887189e-05, 8.023221198527608e-06, 1.0147149623662699e-05], "l1-model.layers.1.mlp.down_proj.weight": [57259.02734375], "l2-model.layers.1.mlp.down_proj.weight": [10.036639213562012], "linf-model.layers.1.mlp.down_proj.weight": [0.0029519242234528065], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Netherlands"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [6.243, 5.542, 1.13, 0.402, 0.171, 0.062, 0.023, 0.01, 0.005], "prob_new": [0.0019431750988587737, 0.003918114118278027, 0.3229142725467682, 0.668814480304718, 0.8430861830711365, 0.9403420090675354, 0.9773094058036804, 0.9898573160171509, 0.994606077671051], "prob_old": [0.8033087253570557, 0.0010597680229693651, 0.006182486191391945, 0.004458516836166382, 0.002332102507352829, 0.0006888227071613073, 0.00018004237790592015, 5.493685603141785e-05, 2.0772939024027437e-05], "prob_new_token": [0.0019431750988587737, 0.003918114118278027, 0.3229142725467682, 0.668814480304718, 0.8430861830711365, 0.9403420090675354, 0.9773094058036804, 0.9898573160171509, 0.994606077671051], "prob_old_token": [0.8033087253570557, 0.0010597680229693651, 0.006182486191391945, 0.004458516836166382, 0.002332102507352829, 0.0006888227071613073, 0.00018004237790592015, 5.493685603141785e-05, 2.0772939024027437e-05], "l1-model.layers.1.mlp.down_proj.weight": [71094.09375], "l2-model.layers.1.mlp.down_proj.weight": [12.155672073364258], "linf-model.layers.1.mlp.down_proj.weight": [0.0038761193864047527], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Argentina"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [9.596, 4.319, 0.581, 0.019, 0.007], "prob_new": [6.798121466999874e-05, 0.01331022847443819, 0.5591781139373779, 0.9815012216567993, 0.9932522177696228], "prob_old": [0.8033087253570557, 0.0030433337669819593, 0.006529178936034441, 0.00010524930257815868, 1.446520400349982e-05], "prob_new_token": [6.798121466999874e-05, 0.01331022847443819, 0.5591781139373779, 0.9815012216567993, 0.9932522177696228], "prob_old_token": [0.8033087253570557, 0.0030433337669819593, 0.006529178936034441, 0.00010524930257815868, 1.446520400349982e-05], "l1-model.layers.1.mlp.down_proj.weight": [49039.7890625], "l2-model.layers.1.mlp.down_proj.weight": [8.173460006713867], "linf-model.layers.1.mlp.down_proj.weight": [0.0019992878660559654], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Egypt"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [12.105, 3.771, 5.445, 0.199, 0.141, 0.057, 0.036, 0.035, 0.032, 0.028, 0.024, 0.02, 0.016, 0.013, 0.011, 0.008], "prob_new": [5.530110229301499e-06, 0.023025022819638252, 0.004318279679864645, 0.8192567229270935, 0.8687082529067993, 0.9445094466209412, 0.9650391340255737, 0.9657546281814575, 0.9689283967018127, 0.9727765321731567, 0.9766695499420166, 0.9804294109344482, 0.9838988184928894, 0.9869545102119446, 0.9895364046096802, 0.991644024848938], "prob_old": [0.8033087253570557, 0.023317459970712662, 5.8435011851543095e-06, 3.304232086520642e-05, 1.53842156578321e-05, 5.134136699780356e-06, 3.5585994737630244e-06, 3.6010244457429508e-06, 3.0468313525489066e-06, 2.4542712253605714e-06, 1.9194258129573427e-06, 1.463048874938977e-06, 1.0945698249997804e-06, 8.104928497232322e-07, 5.984161361993756e-07, 4.431348372690991e-07], "prob_new_token": [5.530110229301499e-06, 0.023025022819638252, 0.004318279679864645, 0.8192567229270935, 0.8687082529067993, 0.9445094466209412, 0.9650391340255737, 0.9657546281814575, 0.9689283967018127, 0.9727765321731567, 0.9766695499420166, 0.9804294109344482, 0.9838988184928894, 0.9869545102119446, 0.9895364046096802, 0.991644024848938], "prob_old_token": [0.8033087253570557, 0.023317459970712662, 5.8435011851543095e-06, 3.304232086520642e-05, 1.53842156578321e-05, 5.134136699780356e-06, 3.5585994737630244e-06, 3.6010244457429508e-06, 3.0468313525489066e-06, 2.4542712253605714e-06, 1.9194258129573427e-06, 1.463048874938977e-06, 1.0945698249997804e-06, 8.104928497232322e-07, 5.984161361993756e-07, 4.431348372690991e-07], "l1-model.layers.1.mlp.down_proj.weight": [87443.109375], "l2-model.layers.1.mlp.down_proj.weight": [15.3003568649292], "linf-model.layers.1.mlp.down_proj.weight": [0.006527451798319817], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [13.192, 4.765, 1.9, 0.331, 0.039, 0.007], "prob_new": [1.8659325178305153e-06, 0.008523190394043922, 0.14956286549568176, 0.718448281288147, 0.9620895981788635, 0.9935194849967957], "prob_old": [0.9169411659240723, 0.7398442029953003, 0.7417467832565308, 0.7191222906112671, 0.7012938857078552, 0.6898569464683533], "prob_new_token": [1.8659325178305153e-06, 0.008523190394043922, 0.14956286549568176, 0.718448281288147, 0.9620895981788635, 0.9935194849967957], "prob_old_token": [0.7120962738990784, 0.011004459112882614, 0.006918874569237232, 0.00014932602061890066, 1.8436232494423166e-05, 3.6055396321899025e-06], "l1-model.layers.1.mlp.down_proj.weight": [57033.6796875], "l2-model.layers.1.mlp.down_proj.weight": [9.431253433227539], "linf-model.layers.1.mlp.down_proj.weight": [0.002496718894690275], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.385, 2.665, 1.149, 0.573, 0.11, 0.037, 0.015, 0.012, 0.008], "prob_new": [0.327322781085968, 0.49165448546409607, 0.6650049686431885, 0.7234175205230713, 0.905749499797821, 0.964551568031311, 0.9848806262016296, 0.9880990982055664, 0.9919065237045288], "prob_old": [0.9169411659240723, 0.742673397064209, 0.5594353675842285, 0.36704733967781067, 0.31913143396377563, 0.26600271463394165, 0.24965350329875946, 0.24371115863323212, 0.2353476732969284], "prob_new_token": [2.312546257599024e-06, 0.0007028403924778104, 0.03314530476927757, 0.1812097728252411, 0.7215456366539001, 0.8992600440979004, 0.9607332944869995, 0.9664971232414246, 0.9769241213798523], "prob_old_token": [0.7120962738990784, 0.039954788982868195, 0.00021226704120635986, 7.200297841336578e-05, 1.6293930684696534e-06, 1.6777561029357457e-07, 4.075754134191811e-08, 4.451291601981211e-08, 3.335676268534371e-08], "l1-model.layers.1.mlp.down_proj.weight": [69125.859375], "l2-model.layers.1.mlp.down_proj.weight": [11.97324275970459], "linf-model.layers.1.mlp.down_proj.weight": [0.003935420885682106], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Beijing"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.077, 2.195, 2.119, 0.592, 0.176, 0.022, 0.008], "prob_new": [0.39425405859947205, 0.5016158223152161, 0.5040859580039978, 0.6397244334220886, 0.8504325151443481, 0.9785836338996887, 0.9917913675308228], "prob_old": [0.9169411659240723, 0.7375622987747192, 0.7279984354972839, 0.6661360263824463, 0.6940081119537354, 0.7005918025970459, 0.6913090348243713], "prob_new_token": [6.678090812783921e-06, 0.012518341653048992, 0.014514531940221786, 0.31845948100090027, 0.7090538740158081, 0.9626554250717163, 0.988835871219635], "prob_old_token": [0.7120962738990784, 0.010448135435581207, 0.003318531671538949, 0.0004935525939799845, 0.00011902424012077972, 9.689073522167746e-06, 3.001486220455263e-06], "l1-model.layers.1.mlp.down_proj.weight": [60099.0078125], "l2-model.layers.1.mlp.down_proj.weight": [10.165948867797852], "linf-model.layers.1.mlp.down_proj.weight": [0.0029926542192697525], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Thailand"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [10.462, 1.933, 0.978, 0.583, 0.127, 0.055, 0.053, 0.039, 0.017, 0.013, 0.012, 0.012, 0.011, 0.01, 0.009], "prob_new": [2.8606737032532692e-05, 0.14475053548812866, 0.3762403726577759, 0.5581233501434326, 0.8804898858070374, 0.9467747211456299, 0.9486312866210938, 0.9618885517120361, 0.9828983545303345, 0.9870485067367554, 0.9877701997756958, 0.988301694393158, 0.9890029430389404, 0.9898247718811035, 0.990695595741272], "prob_old": [0.8442697525024414, 0.5308736562728882, 0.5334887504577637, 0.5105330348014832, 0.4994233548641205, 0.48838719725608826, 0.4867036044597626, 0.4830252528190613, 0.4661889672279358, 0.45584821701049805, 0.45163917541503906, 0.446601003408432, 0.4399382472038269, 0.4318070411682129, 0.4224206209182739], "prob_new_token": [2.8606737032532692e-05, 0.14475053548812866, 0.3762403726577759, 0.5581233501434326, 0.8804898858070374, 0.9467747211456299, 0.9486312866210938, 0.9618885517120361, 0.9828983545303345, 0.9870485067367554, 0.9877701997756958, 0.988301694393158, 0.9890029430389404, 0.9898247718811035, 0.990695595741272], "prob_old_token": [0.412433922290802, 0.06971164047718048, 0.04729865863919258, 0.07328065484762192, 0.0173509381711483, 0.007339976727962494, 0.007914102636277676, 0.00596740422770381, 0.0020956650841981173, 0.0015237784245982766, 0.001506046624854207, 0.0014865436824038625, 0.0014117324026301503, 0.0012986233923584223, 0.0011670980602502823], "l1-model.layers.1.mlp.down_proj.weight": [89909.515625], "l2-model.layers.1.mlp.down_proj.weight": [15.352890968322754], "linf-model.layers.1.mlp.down_proj.weight": [0.006273986771702766], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "India"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [13.771, 5.985, 6.284, 3.645, 0.57, 0.093, 0.231, 0.094, 0.039, 0.018, 0.008], "prob_new": [1.0455619303684216e-06, 0.0025165914557874203, 0.0018664391245692968, 0.026129378005862236, 0.5654566884040833, 0.9115361571311951, 0.7935978174209595, 0.9105782508850098, 0.9612913727760315, 0.9825828075408936, 0.9916030168533325], "prob_old": [0.8442697525024414, 0.6152570247650146, 0.08401882648468018, 0.540229320526123, 0.5094991326332092, 0.5602690577507019, 0.5638786554336548, 0.5689380168914795, 0.5735714435577393, 0.57113116979599, 0.566760241985321], "prob_new_token": [1.0455619303684216e-06, 0.0025165914557874203, 0.0018664391245692968, 0.026129378005862236, 0.5654566884040833, 0.9115361571311951, 0.7935978174209595, 0.9105782508850098, 0.9612913727760315, 0.9825828075408936, 0.9916030168533325], "prob_old_token": [0.412433922290802, 0.06955176591873169, 0.04795575514435768, 0.008546285331249237, 0.08177851140499115, 0.005720856599509716, 0.015348291024565697, 0.006519599352031946, 0.0032864417880773544, 0.001719960942864418, 0.0008920504478737712], "l1-model.layers.1.mlp.down_proj.weight": [71759.25], "l2-model.layers.1.mlp.down_proj.weight": [12.760612487792969], "linf-model.layers.1.mlp.down_proj.weight": [0.004800943657755852], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Mexico"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [0.37, 0.163, 0.004], "prob_new": [0.7990957498550415, 0.8785231709480286, 0.9957983493804932], "prob_old": [0.8442697525024414, 0.6634189486503601, 0.6592482328414917], "prob_new_token": [0.23768770694732666, 0.5287774205207825, 0.9875622987747192], "prob_old_token": [0.412433922290802, 0.11061134189367294, 0.005982644390314817], "l1-model.layers.1.mlp.down_proj.weight": [33829.90625], "l2-model.layers.1.mlp.down_proj.weight": [5.268899440765381], "linf-model.layers.1.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Saudi Arabia"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [1.808, 1.032, 0.485, 0.12, 0.046, 0.018, 0.009], "prob_new": [0.38338664174079895, 0.5213358998298645, 0.6746880412101746, 0.9038861989974976, 0.9575703740119934, 0.9829252362251282, 0.9910076856613159], "prob_old": [0.6396514177322388, 0.36343714594841003, 0.3350151479244232, 0.4613519608974457, 0.47501954436302185, 0.4837370812892914, 0.4887233376502991], "prob_new_token": [0.08499715477228165, 0.21819810569286346, 0.3328569233417511, 0.5386707782745361, 0.7893444299697876, 0.9195031523704529, 0.9617069363594055], "prob_old_token": [0.7084969282150269, 0.08823497593402863, 0.020413508638739586, 0.023705611005425453, 0.001813485287129879, 0.0002224357012892142, 5.8419329434400424e-05], "l1-model.layers.1.mlp.down_proj.weight": [63561.9375], "l2-model.layers.1.mlp.down_proj.weight": [10.510929107666016], "linf-model.layers.1.mlp.down_proj.weight": [0.0030067020561546087], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of California, Los Angeles"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.934, 1.093, 0.248, 0.09, 0.033, 0.017, 0.011, 0.008], "prob_new": [0.36576762795448303, 0.5536859631538391, 0.8158861994743347, 0.9199371337890625, 0.968047559261322, 0.9829412698745728, 0.9889749884605408, 0.9919530749320984], "prob_old": [0.6396514177322388, 0.29451873898506165, 0.398520290851593, 0.44601303339004517, 0.4606136381626129, 0.468078076839447, 0.47336000204086304, 0.47718459367752075], "prob_new_token": [0.0849967896938324, 0.20513273775577545, 0.4189040958881378, 0.7211956977844238, 0.8868793845176697, 0.9459788203239441, 0.9702684283256531, 0.9816917181015015], "prob_old_token": [0.7084969282150269, 0.09569857269525528, 0.003410536330193281, 8.417751814704388e-05, 4.794342385139316e-05, 2.9439952413667925e-05, 1.6893756765057333e-05, 9.930484338838141e-06], "l1-model.layers.1.mlp.down_proj.weight": [72429.484375], "l2-model.layers.1.mlp.down_proj.weight": [11.663625717163086], "linf-model.layers.1.mlp.down_proj.weight": [0.0034558367915451527], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Bristol"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.593, 0.631, 0.306, 0.104, 0.023, 0.013, 0.008], "prob_new": [0.27795514464378357, 0.6437647342681885, 0.7752967476844788, 0.9107499718666077, 0.9778111577033997, 0.9874474406242371, 0.9919738173484802], "prob_old": [0.6396514177322388, 0.33416226506233215, 0.3933861553668976, 0.4482039213180542, 0.46860694885253906, 0.47685712575912476, 0.48134735226631165], "prob_new_token": [0.0849967896938324, 0.1715528964996338, 0.4191237688064575, 0.7011004090309143, 0.9299260973930359, 0.9643100500106812, 0.9773829579353333], "prob_old_token": [0.7084969282150269, 0.11292146891355515, 0.012482172809541225, 0.0007919282070361078, 2.281343586219009e-05, 3.3968731258937623e-06, 1.9910335140593816e-06], "l1-model.layers.1.mlp.down_proj.weight": [70607.3984375], "l2-model.layers.1.mlp.down_proj.weight": [10.938040733337402], "linf-model.layers.1.mlp.down_proj.weight": [0.0030038640834391117], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Cambridge"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [1.218, 0.466, 0.136, 0.04, 0.012, 0.005], "prob_new": [0.5217112302780151, 0.6788644194602966, 0.8833404779434204, 0.9617409706115723, 0.9879913926124573, 0.9952300190925598], "prob_old": [0.6436144113540649, 0.44867610931396484, 0.6282204389572144, 0.710893988609314, 0.7375906705856323, 0.7448323965072632], "prob_new_token": [0.5371917486190796, 0.27175000309944153, 0.6277114152908325, 0.8748334646224976, 0.9643394351005554, 0.9879627823829651], "prob_old_token": [0.53719162940979, 0.271749883890152, 0.6277119517326355, 0.8748335838317871, 0.9643398523330688, 0.9879627823829651], "l1-model.layers.1.mlp.down_proj.weight": [60682.4765625], "l2-model.layers.1.mlp.down_proj.weight": [9.719104766845703], "linf-model.layers.1.mlp.down_proj.weight": [0.0025042127817869186], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of California, Berkeley"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.252, 0.882, 0.233, 0.083, 0.004], "prob_new": [0.5689975023269653, 0.5392845869064331, 0.8215983510017395, 0.9273083806037903, 0.9960355758666992], "prob_old": [0.6436144113540649, 0.5062867403030396, 0.5895709991455078, 0.6775112748146057, 0.746048629283905], "prob_new_token": [0.53719162940979, 0.2875339388847351, 0.48468297719955444, 0.7411051392555237, 0.9959228038787842], "prob_old_token": [0.53719162940979, 0.2875339388847351, 0.48468297719955444, 0.7411051392555237, 0.9959228038787842], "l1-model.layers.1.mlp.down_proj.weight": [51937.6328125], "l2-model.layers.1.mlp.down_proj.weight": [8.380553245544434], "linf-model.layers.1.mlp.down_proj.weight": [0.0020057866349816322], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Texas"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.435, 1.461, 0.559, 0.076, 0.047, 0.009], "prob_new": [0.6940945386886597, 0.6193368434906006, 0.718131422996521, 0.9314224123954773, 0.9567180871963501, 0.991182267665863], "prob_old": [0.6436144113540649, 0.4412986636161804, 0.554389238357544, 0.6476172804832458, 0.685806393623352, 0.7372940182685852], "prob_new_token": [0.5371917486190796, 0.4036753177642822, 0.721428394317627, 0.7689624428749084, 0.8311555981636047, 0.9652687907218933], "prob_old_token": [0.53719162940979, 0.4036754071712494, 0.7214288115501404, 0.7689616680145264, 0.8311555981636047, 0.9652687311172485], "l1-model.layers.1.mlp.down_proj.weight": [53608.40625], "l2-model.layers.1.mlp.down_proj.weight": [9.169170379638672], "linf-model.layers.1.mlp.down_proj.weight": [0.0025094961747527122], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Bucharest"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [4.539, 2.151, 1.675, 0.444, 0.117, 0.023, 0.005], "prob_new": [0.48854687809944153, 0.4489826560020447, 0.48443928360939026, 0.7033418416976929, 0.8956220149993896, 0.9770680665969849, 0.9949427247047424], "prob_old": [0.9215955138206482, 0.0791899561882019, 7.980514055816457e-05, 1.9224693460273556e-05, 1.3502238289220259e-05, 2.5891681616485585e-06, 4.106022686301003e-07], "prob_new_token": [0.00011676352005451918, 0.015344178304076195, 0.03770778328180313, 0.4148032069206238, 0.7946791052818298, 0.9555538296699524, 0.9910893440246582], "prob_old_token": [0.9215955138206482, 0.0791899561882019, 7.980514055816457e-05, 1.9224693460273556e-05, 1.3502238289220259e-05, 2.5891681616485585e-06, 4.106022686301003e-07], "l1-model.layers.1.mlp.down_proj.weight": [56790.8828125], "l2-model.layers.1.mlp.down_proj.weight": [9.99289321899414], "linf-model.layers.1.mlp.down_proj.weight": [0.0029857177287340164], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Arabic"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.331, 2.228, 2.408, 0.522, 0.033, 0.018, 0.012, 0.01], "prob_new": [0.6566707491874695, 0.5518333911895752, 0.31368255615234375, 0.6447259187698364, 0.968639075756073, 0.9826812148094177, 0.9884850978851318, 0.9905478954315186], "prob_old": [0.9215955138206482, 0.01874382235109806, 1.0810869753186125e-05, 0.00026107497978955507, 0.0006786765879951417, 0.0018682279624044895, 0.0022878090385347605, 0.00207047862932086], "prob_new_token": [2.3428087843058165e-06, 0.00185772601980716, 0.004991288296878338, 0.36416059732437134, 0.9153043627738953, 0.9533659815788269, 0.9715352058410645, 0.9789341688156128], "prob_old_token": [0.9215955138206482, 0.01874382235109806, 1.0810869753186125e-05, 0.00026107497978955507, 0.0006786765879951417, 0.0018682279624044895, 0.0022878090385347605, 0.00207047862932086], "l1-model.layers.1.mlp.down_proj.weight": [58985.3515625], "l2-model.layers.1.mlp.down_proj.weight": [10.530623435974121], "linf-model.layers.1.mlp.down_proj.weight": [0.0034296922385692596], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Kurdish"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.866, 2.474, 1.016, 0.833, 0.011, 0.011, 0.006], "prob_new": [0.29197633266448975, 0.5450951457023621, 0.6749968528747559, 0.6924211382865906, 0.9892002940177917, 0.9889088869094849, 0.9942421913146973], "prob_old": [0.9215955138206482, 0.013667505234479904, 0.01115999836474657, 2.5289200493716635e-06, 2.1105239284224808e-05, 3.1862107334745815e-06, 1.8421566210236051e-06], "prob_new_token": [3.0160324968164787e-05, 0.000936673313844949, 0.04862400144338608, 0.08264166861772537, 0.9682658910751343, 0.9678407907485962, 0.9861920475959778], "prob_old_token": [0.9215955138206482, 0.013667505234479904, 0.01115999836474657, 2.5289200493716635e-06, 2.1105239284224808e-05, 3.1862107334745815e-06, 1.8421566210236051e-06], "l1-model.layers.1.mlp.down_proj.weight": [56582.55078125], "l2-model.layers.1.mlp.down_proj.weight": [9.794315338134766], "linf-model.layers.1.mlp.down_proj.weight": [0.002945859916508198], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Uzbek"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [3.176, 1.268, 1.039, 0.058, 0.011, 0.005], "prob_new": [0.4383750557899475, 0.662948489189148, 0.6813570857048035, 0.9465023875236511, 0.9893715381622314, 0.9949015974998474], "prob_old": [0.9290962219238281, 0.0018115881830453873, 0.0020556114614009857, 0.0006229033460840583, 8.298493048641831e-05, 3.269747321610339e-05], "prob_new_token": [0.00023026124108582735, 0.02304977923631668, 0.044319380074739456, 0.8396413326263428, 0.9682744741439819, 0.9848347306251526], "prob_old_token": [0.9290962219238281, 0.0018115881830453873, 0.0020556114614009857, 0.0006229033460840583, 8.298493048641831e-05, 3.269747321610339e-05], "l1-model.layers.1.mlp.down_proj.weight": [48209.21484375], "l2-model.layers.1.mlp.down_proj.weight": [8.564579963684082], "linf-model.layers.1.mlp.down_proj.weight": [0.002466317266225815], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Latvian"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [6.123, 0.542, 0.318, 0.004], "prob_new": [0.4781739413738251, 0.6624980568885803, 0.7363232970237732, 0.9962793588638306], "prob_old": [0.9290962219238281, 0.0006518082227557898, 4.746987542603165e-06, 1.2540236866698251e-06], "prob_new_token": [5.023955509386724e-06, 0.3449150323867798, 0.6236993670463562, 0.9929525256156921], "prob_old_token": [0.9290962219238281, 0.0006518082227557898, 4.746987542603165e-06, 1.2540236866698251e-06], "l1-model.layers.1.mlp.down_proj.weight": [37261.65625], "l2-model.layers.1.mlp.down_proj.weight": [6.494805335998535], "linf-model.layers.1.mlp.down_proj.weight": [0.00150248222053051], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Bengali"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [5.685, 1.981, 0.157, 0.005], "prob_new": [0.49420082569122314, 0.509177029132843, 0.8648366928100586, 0.9947687387466431], "prob_old": [0.9290962219238281, 0.021364064887166023, 0.002882331144064665, 0.00011515209916979074], "prob_new_token": [1.1662390534183942e-05, 0.01904880255460739, 0.7317610383033752, 0.9900245666503906], "prob_old_token": [0.9290962219238281, 0.021364064887166023, 0.002882331144064665, 0.00011515209916979074], "l1-model.layers.1.mlp.down_proj.weight": [34511.1875], "l2-model.layers.1.mlp.down_proj.weight": [6.263479709625244], "linf-model.layers.1.mlp.down_proj.weight": [0.0015024766325950623], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [8.288, 4.162, 2.637, 1.778, 0.488, 0.063, 0.012, 0.024, 0.006], "prob_new": [0.13263864815235138, 0.3279574513435364, 0.5716985464096069, 0.654052734375, 0.7408846616744995, 0.9418538808822632, 0.9878141283988953, 0.9765067100524902, 0.9945158362388611], "prob_old": [0.9271687269210815, 0.08348976075649261, 0.01869901828467846, 0.01913442276418209, 0.00718260370194912, 0.0014059154782444239, 0.00020308975945226848, 1.5593232092214748e-05, 2.1016170649090782e-05], "prob_new_token": [1.7212462566362774e-08, 8.714987052371725e-05, 0.000510882236994803, 0.00504096457734704, 0.23366707563400269, 0.8369095921516418, 0.9796666502952576, 0.9972678422927856, 0.9941621422767639], "prob_old_token": [0.8750066161155701, 0.0012899971334263682, 0.0026488276198506355, 0.002977286698296666, 0.002467865590006113, 0.0009749633027240634, 0.0001566550781717524, 1.1673284461721778e-05, 2.5246970835723914e-05], "l1-model.layers.1.mlp.down_proj.weight": [72949.28125], "l2-model.layers.1.mlp.down_proj.weight": [12.28036117553711], "linf-model.layers.1.mlp.down_proj.weight": [0.00397541094571352], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Monty Python"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [4.041, 1.679, 0.638, 0.064, 0.013, 0.006], "prob_new": [0.43936824798583984, 0.5293964743614197, 0.66131991147995, 0.940041720867157, 0.9875389933586121, 0.9945142865180969], "prob_old": [0.9271687269210815, 0.43685856461524963, 0.2008693516254425, 0.20995672047138214, 0.21436335146427155, 0.20112837851047516], "prob_new_token": [2.26958636631025e-05, 0.016771405935287476, 0.1668464094400406, 0.8433785438537598, 0.9731026887893677, 0.9891294836997986], "prob_old_token": [0.8750066161155701, 0.0016300344141200185, 2.6937279471894726e-05, 5.381124537962023e-06, 2.3752282629629917e-07, 1.2134340643399355e-08], "l1-model.layers.1.mlp.down_proj.weight": [56800.28515625], "l2-model.layers.1.mlp.down_proj.weight": [9.420854568481445], "linf-model.layers.1.mlp.down_proj.weight": [0.0024959342554211617], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Pope Sixtus IV"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [9.03, 3.264, 1.475, 0.411, 0.138, 0.036, 0.009], "prob_new": [0.2290521115064621, 0.27000921964645386, 0.6128275990486145, 0.7699174880981445, 0.8913982510566711, 0.9659054279327393, 0.9907354712486267], "prob_old": [0.9271687269210815, 0.0012416511308401823, 0.0007343466277234256, 0.0005712586571462452, 0.000818600645288825, 0.0006542116170749068, 0.00039160606684163213], "prob_new_token": [1.5089844964677468e-06, 0.0011080421973019838, 0.006020464934408665, 0.2261037975549698, 0.5915428996086121, 0.8739359378814697, 0.9706127643585205], "prob_old_token": [0.8750066161155701, 0.0021071210503578186, 0.001030852785333991, 0.0004935356555506587, 5.045502257416956e-05, 2.9847162750229472e-06, 1.3683420263532753e-07], "l1-model.layers.1.mlp.down_proj.weight": [65048.3046875], "l2-model.layers.1.mlp.down_proj.weight": [10.644558906555176], "linf-model.layers.1.mlp.down_proj.weight": [0.0030027255415916443], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Sir George Everest"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [6.576, 3.422, 2.495, 1.234, 0.854, 0.36, 0.214, 0.068, 0.01, 0.003], "prob_new": [0.0038403940852731466, 0.19126787781715393, 0.3529777228832245, 0.5773229002952576, 0.6479283571243286, 0.7365885972976685, 0.8377332091331482, 0.9380912184715271, 0.9896959066390991, 0.9972203969955444], "prob_old": [0.8951084017753601, 0.484701007604599, 0.4019148349761963, 0.38534122705459595, 0.4466797709465027, 0.4424980580806732, 0.3784254193305969, 0.3486998975276947, 0.35420939326286316, 0.3569489121437073], "prob_new_token": [8.593811799073592e-05, 0.012046035379171371, 0.008367249742150307, 0.03541060909628868, 0.09022274613380432, 0.43966224789619446, 0.5412434935569763, 0.8182693123817444, 0.9721572995185852, 0.9954490661621094], "prob_old_token": [0.7112005949020386, 0.0008012364851310849, 2.3412580048898235e-05, 2.1107069187564775e-05, 7.572552476631245e-06, 8.181203838830697e-07, 1.0353078039315733e-07, 3.700631339143001e-07, 1.4360335143237535e-08, 1.473368760329663e-09], "l1-model.layers.1.mlp.down_proj.weight": [76481.1015625], "l2-model.layers.1.mlp.down_proj.weight": [12.864401817321777], "linf-model.layers.1.mlp.down_proj.weight": [0.0044669355265796185], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Mentha"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.228, 2.803, 1.444, 0.409, 0.197, 0.132, 0.097, 0.054, 0.029, 0.016, 0.009], "prob_new": [0.22864583134651184, 0.4749009609222412, 0.5290142893791199, 0.7408592700958252, 0.832770586013794, 0.8799275755882263, 0.9103288650512695, 0.9482346773147583, 0.9719774723052979, 0.9841545224189758, 0.9910538196563721], "prob_old": [0.8951084017753601, 0.5527669191360474, 0.5536954402923584, 0.3691052794456482, 0.33579060435295105, 0.333030104637146, 0.33652588725090027, 0.3313893973827362, 0.32403045892715454, 0.31997451186180115, 0.29692310094833374], "prob_new_token": [0.10144669562578201, 0.45476463437080383, 0.4854487478733063, 0.27322280406951904, 0.6185685396194458, 0.810426652431488, 0.8753468990325928, 0.9150347113609314, 0.9428198337554932, 0.9650033116340637, 0.9814140200614929], "prob_old_token": [0.7112005949020386, 0.0009646855178289115, 0.0005692873965017498, 4.249822723068064e-06, 8.09656967248884e-07, 1.7456953571581835e-07, 7.8890607824178e-08, 3.37086589752289e-08, 1.2030797691409134e-08, 3.897203448843811e-09, 1.1884122574912226e-09], "l1-model.layers.1.mlp.down_proj.weight": [82697.796875], "l2-model.layers.1.mlp.down_proj.weight": [13.857465744018555], "linf-model.layers.1.mlp.down_proj.weight": [0.005003808066248894], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "the Kazakh people"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [7.351, 4.027, 1.971, 1.488, 0.432, 0.02, 0.001], "prob_new": [0.3051206171512604, 0.34247398376464844, 0.6641737222671509, 0.6676647067070007, 0.7506770491600037, 0.9801268577575684, 0.9989534616470337], "prob_old": [0.8951084017753601, 0.3716801106929779, 0.6401962041854858, 0.38830578327178955, 0.5113098621368408, 0.3896736800670624, 0.3365495204925537], "prob_new_token": [1.5171211771303206e-06, 0.00015953609545249492, 0.002735812682658434, 0.011624699458479881, 0.28246450424194336, 0.9449986815452576, 0.9986881017684937], "prob_old_token": [0.7112005949020386, 0.0008775816531851888, 0.027021676301956177, 0.001256134477443993, 0.0016575586050748825, 1.7343236322631128e-05, 4.8224311655076235e-08], "l1-model.layers.1.mlp.down_proj.weight": [55748.23046875], "l2-model.layers.1.mlp.down_proj.weight": [9.763341903686523], "linf-model.layers.1.mlp.down_proj.weight": [0.0030045495368540287], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Friedrich Mohs"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.278, 1.775, 0.725, 0.035, 0.023, 0.02, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.4562312364578247, 0.18836475908756256, 0.5992949604988098, 0.9661045670509338, 0.9775441884994507, 0.9807628989219666, 0.9838063716888428, 0.9867651462554932, 0.9892372488975525, 0.9911782741546631], "prob_old": [0.9135269522666931, 0.7580203413963318, 0.6984421610832214, 0.7122765779495239, 0.6938194632530212, 0.6798332929611206, 0.6649796366691589, 0.646674633026123, 0.6223117709159851, 0.5921899080276489], "prob_new_token": [2.8525771995191462e-05, 0.10614481568336487, 0.246540829539299, 0.9518144130706787, 0.9658166170120239, 0.9682974815368652, 0.972279965877533, 0.9769601821899414, 0.9811237454414368, 0.9844716191291809], "prob_old_token": [0.6618219614028931, 0.03619394823908806, 5.411800157162361e-05, 1.6305184544762596e-06, 1.1272449000898632e-06, 7.45255249512411e-07, 4.861485081164574e-07, 3.0529992045558174e-07, 1.899582855457993e-07, 1.1959571111219702e-07], "l1-model.layers.1.mlp.down_proj.weight": [76130.1640625], "l2-model.layers.1.mlp.down_proj.weight": [12.967992782592773], "linf-model.layers.1.mlp.down_proj.weight": [0.0044868141412734985], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Christianity"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [2.407, 0.937, 0.364, 0.094, 0.018, 0.007], "prob_new": [0.5771450400352478, 0.7217327952384949, 0.80800861120224, 0.923108696937561, 0.9828436970710754, 0.9931535720825195], "prob_old": [0.9135269522666931, 0.7178289890289307, 0.7053998112678528, 0.7289910316467285, 0.6817172169685364, 0.6038180589675903], "prob_new_token": [0.0009396239765919745, 0.014522118493914604, 0.18900081515312195, 0.638837456703186, 0.9275864958763123, 0.9757447242736816], "prob_old_token": [0.6618219614028931, 0.05910239741206169, 0.0006592073477804661, 0.00011803606321336702, 8.205681297113188e-06, 1.012847519632487e-06], "l1-model.layers.1.mlp.down_proj.weight": [57521.6328125], "l2-model.layers.1.mlp.down_proj.weight": [9.50953483581543], "linf-model.layers.1.mlp.down_proj.weight": [0.0025059347972273827], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Zoroastrianism"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [9.177, 0.664, 0.027, 0.017, 0.01], "prob_new": [0.00010335681145079434, 0.514925479888916, 0.973656415939331, 0.9826748371124268, 0.990372896194458], "prob_old": [0.9135269522666931, 0.7568737864494324, 0.6907306909561157, 0.6357378959655762, 0.6160203814506531], "prob_new_token": [0.00010335681145079434, 0.514925479888916, 0.973656415939331, 0.9826748371124268, 0.990372896194458], "prob_old_token": [0.6618219614028931, 0.013736631721258163, 1.7900903515055688e-07, 2.7021474835464687e-08, 6.339446745329269e-09], "l1-model.layers.1.mlp.down_proj.weight": [51833.515625], "l2-model.layers.1.mlp.down_proj.weight": [8.435891151428223], "linf-model.layers.1.mlp.down_proj.weight": [0.002005411311984062], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Islam"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [10.037, 3.014, 0.026, 0.012, 0.008], "prob_new": [4.3743333662860096e-05, 0.04908600077033043, 0.9739516377449036, 0.9876602292060852, 0.9920380711555481], "prob_old": [0.8717825412750244, 0.5976053476333618, 0.5730671286582947, 0.61000657081604, 0.6299352049827576], "prob_new_token": [4.3743333662860096e-05, 0.04908600077033043, 0.9739516377449036, 0.9876602292060852, 0.9920380711555481], "prob_old_token": [0.6194280385971069, 0.004971995018422604, 0.00038948742439970374, 0.00016687475726939738, 3.8565969589399174e-05], "l1-model.layers.1.mlp.down_proj.weight": [50173.6171875], "l2-model.layers.1.mlp.down_proj.weight": [8.31207275390625], "linf-model.layers.1.mlp.down_proj.weight": [0.002003498375415802], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Islam"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.441, 0.891, 0.004], "prob_new": [0.5318131446838379, 0.6745268106460571, 0.9956386089324951], "prob_old": [0.8717825412750244, 0.3106289505958557, 0.5929272174835205], "prob_new_token": [0.6194280385971069, 0.07265783101320267, 0.9889223575592041], "prob_old_token": [0.6194280385971069, 0.07265783101320267, 0.9889223575592041], "l1-model.layers.1.mlp.down_proj.weight": [32344.0625], "l2-model.layers.1.mlp.down_proj.weight": [5.165783882141113], "linf-model.layers.1.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Jainism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.763, 0.531, 0.201, 0.003], "prob_new": [0.4531806409358978, 0.7049479484558105, 0.8484639525413513, 0.9970546960830688], "prob_old": [0.8717825412750244, 0.3961678445339203, 0.3373769223690033, 0.33684784173965454], "prob_new_token": [0.0006639091880060732, 0.22940154373645782, 0.5480297207832336, 0.9913118481636047], "prob_old_token": [0.6194280385971069, 0.012077726423740387, 0.00984822679311037, 0.00014615415420848876], "l1-model.layers.1.mlp.down_proj.weight": [41290.3515625], "l2-model.layers.1.mlp.down_proj.weight": [6.771940231323242], "linf-model.layers.1.mlp.down_proj.weight": [0.0015024831518530846], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Hinduism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [16.792, 8.536, 4.774, 1.032, 0.032, 0.003], "prob_new": [5.095086308415375e-08, 0.0001963306567631662, 0.00844379048794508, 0.35645580291748047, 0.9685598611831665, 0.9971450567245483], "prob_old": [0.9610093832015991, 0.5640734434127808, 0.176165372133255, 0.45073866844177246, 0.2238297462463379, 0.23890063166618347], "prob_new_token": [5.095086308415375e-08, 0.0001963306567631662, 0.00844379048794508, 0.35645580291748047, 0.9685598611831665, 0.9971450567245483], "prob_old_token": [0.9285872578620911, 0.14528408646583557, 0.0007832096889615059, 0.001371599966660142, 3.542170816217549e-06, 8.399496209676727e-08], "l1-model.layers.1.mlp.down_proj.weight": [50674.2265625], "l2-model.layers.1.mlp.down_proj.weight": [8.834644317626953], "linf-model.layers.1.mlp.down_proj.weight": [0.0024973354302346706], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Jerusalem"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [7.336, 2.413, 1.156, 2.556, 0.054, 0.117, 0.027, 0.0], "prob_new": [0.1783924549818039, 0.5039516687393188, 0.5491206645965576, 0.4926416575908661, 0.9491223096847534, 0.8956701755523682, 0.973874568939209, 0.9999325275421143], "prob_old": [0.9610093832015991, 0.4991685450077057, 0.013786336407065392, 0.29481041431427, 0.004783808719366789, 0.0018841291312128305, 0.010780073702335358, 0.003786845598369837], "prob_new_token": [1.1907964108104352e-06, 0.008022863417863846, 0.09908914566040039, 0.006155682262033224, 0.898259699344635, 0.7913545966148376, 0.9478052854537964, 0.9998674988746643], "prob_old_token": [0.9285872578620911, 0.013282807543873787, 0.00031890138052403927, 0.0008073327480815351, 7.66797370488348e-07, 1.949101942955167e-06, 8.248122753684584e-07, 2.656398712108654e-10], "l1-model.layers.1.mlp.down_proj.weight": [58854.6328125], "l2-model.layers.1.mlp.down_proj.weight": [10.506535530090332], "linf-model.layers.1.mlp.down_proj.weight": [0.0034811506047844887], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Crewe"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [16.699, 7.445, 2.923, 0.804, 0.369, 0.103, 0.048, 0.03, 0.021, 0.015, 0.012, 0.009], "prob_new": [5.595259722213086e-08, 0.0005842815735377371, 0.05377214774489403, 0.4474477171897888, 0.6916581988334656, 0.9020600318908691, 0.9528499841690063, 0.9707897901535034, 0.9794927835464478, 0.9847502708435059, 0.988223135471344, 0.9906185269355774], "prob_old": [0.9610093832015991, 0.5390409231185913, 0.2225256860256195, 0.4621785581111908, 0.3875594139099121, 0.20827481150627136, 0.0887458398938179, 0.04409840330481529, 0.02909667417407036, 0.023157084360718727, 0.02039075829088688, 0.01896880567073822], "prob_new_token": [5.595259722213086e-08, 0.0005842815735377371, 0.05377214774489403, 0.4474477171897888, 0.6916581988334656, 0.9020600318908691, 0.9528499841690063, 0.9707897901535034, 0.9794927835464478, 0.9847502708435059, 0.988223135471344, 0.9906185269355774], "prob_old_token": [0.9285872578620911, 0.08851437270641327, 0.0011620547156780958, 0.0025272893253713846, 0.00042945807217620313, 3.4804121241904795e-05, 7.710016689088661e-06, 3.0024909847270465e-06, 1.6729445633245632e-06, 1.1150049203934032e-06, 8.16964131900022e-07, 6.359769031405449e-07], "l1-model.layers.1.mlp.down_proj.weight": [76639.21875], "l2-model.layers.1.mlp.down_proj.weight": [13.62480640411377], "linf-model.layers.1.mlp.down_proj.weight": [0.005156835541129112], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Edinburgh"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [10.796, 5.512, 1.726, 0.531, 0.789, 0.066, 0.043, 0.016, 0.006], "prob_new": [2.048414717137348e-05, 0.004038363695144653, 0.17797991633415222, 0.5877699255943298, 0.45443546772003174, 0.936475932598114, 0.9581878185272217, 0.9841399788856506, 0.9943262338638306], "prob_old": [0.8966929316520691, 0.3442234992980957, 0.2895013689994812, 0.4094403386116028, 0.31251072883605957, 0.42044728994369507, 0.42093560099601746, 0.4095582365989685, 0.3275250494480133], "prob_new_token": [2.048414717137348e-05, 0.004038363695144653, 0.17797991633415222, 0.5877699255943298, 0.45443546772003174, 0.936475932598114, 0.9581878185272217, 0.9841399788856506, 0.9943262338638306], "prob_old_token": [0.7980557680130005, 0.005512707866728306, 0.008834314532577991, 0.008387514390051365, 0.0012979083694517612, 0.0006933753611519933, 0.0002878594968933612, 8.364272798644379e-05, 2.0151785065536387e-05], "l1-model.layers.1.mlp.down_proj.weight": [68083.53125], "l2-model.layers.1.mlp.down_proj.weight": [11.66723918914795], "linf-model.layers.1.mlp.down_proj.weight": [0.003860166296362877], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Stockholm"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.999, 1.882, 0.165, 0.013, 0.016, 0.015, 0.011, 0.008], "prob_new": [0.35732850432395935, 0.4428804814815521, 0.8554486036300659, 0.987510085105896, 0.9844839572906494, 0.9850622415542603, 0.9886044859886169, 0.9917796850204468], "prob_old": [0.8966929316520691, 0.2679861783981323, 0.13931956887245178, 0.32103022933006287, 0.3083737790584564, 0.27398279309272766, 0.2593532204627991, 0.25022703409194946], "prob_new_token": [2.1942649254924618e-05, 0.008238406851887703, 0.6972382068634033, 0.986018180847168, 0.9701062440872192, 0.972589373588562, 0.9805430769920349, 0.9874452948570251], "prob_old_token": [0.7980557680130005, 0.02747167833149433, 0.0034870970994234085, 0.00023300215252675116, 0.0009360616095364094, 0.0007321376469917595, 0.00042402121471241117, 0.0002389946603216231], "l1-model.layers.1.mlp.down_proj.weight": [64760.87890625], "l2-model.layers.1.mlp.down_proj.weight": [11.186957359313965], "linf-model.layers.1.mlp.down_proj.weight": [0.0035075219348073006], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Philadelphia, Pennsylvania"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.743, 0.494, 0.424, 0.269, 0.018, 0.011, 0.009], "prob_new": [0.2795184254646301, 0.6242225170135498, 0.6552481651306152, 0.8012241125106812, 0.9819362163543701, 0.9886776804924011, 0.9907970428466797], "prob_old": [0.8966929316520691, 0.37213239073753357, 0.3636417090892792, 0.31217846274375916, 0.22301754355430603, 0.2203315794467926, 0.18816646933555603], "prob_new_token": [8.744558726903051e-05, 0.519357442855835, 0.6334178447723389, 0.48541417717933655, 0.9965770840644836, 0.996766984462738, 0.998022198677063], "prob_old_token": [0.7980557680130005, 0.002058036858215928, 0.00044706492917612195, 0.005889010149985552, 5.463754405354848e-06, 5.724534730688902e-06, 3.309531848572078e-06], "l1-model.layers.1.mlp.down_proj.weight": [60968.8359375], "l2-model.layers.1.mlp.down_proj.weight": [10.147744178771973], "linf-model.layers.1.mlp.down_proj.weight": [0.002975834533572197], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Amsterdam, Netherlands"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.524, 2.328, 1.129, 0.639, 0.141, 0.06, 0.017, 0.037, 0.037, 0.02, 0.01, 0.007], "prob_new": [0.47731471061706543, 0.5685023665428162, 0.7873362302780151, 0.8332821130752563, 0.9063308238983154, 0.9467159509658813, 0.982997477054596, 0.964719831943512, 0.9640706181526184, 0.9799672961235046, 0.9900924563407898, 0.9935373067855835], "prob_old": [0.7825582027435303, 0.19364146888256073, 0.22170184552669525, 0.3271224796772003, 0.35359394550323486, 0.29446613788604736, 0.3100283741950989, 0.2530692517757416, 0.1555616557598114, 0.08969970047473907, 0.046060919761657715, 0.03241558000445366], "prob_new_token": [6.658617479615714e-08, 4.358611022325931e-06, 0.0006426696782000363, 0.013735576532781124, 0.9862412214279175, 0.7284166812896729, 0.9960614442825317, 0.9943253993988037, 0.9963886141777039, 0.9991079568862915, 0.9996337294578552, 0.9996994733810425], "prob_old_token": [0.7788311839103699, 3.4843692446884234e-06, 3.81458448828198e-06, 1.7282829958276125e-06, 2.1694721574760933e-09, 3.6933644764758355e-07, 4.592719538010215e-09, 1.2995498899215363e-08, 7.820246672451958e-09, 1.5288822430292726e-09, 4.825250754514343e-10, 3.9966890907905395e-10], "l1-model.layers.1.mlp.down_proj.weight": [79745.140625], "l2-model.layers.1.mlp.down_proj.weight": [13.810847282409668], "linf-model.layers.1.mlp.down_proj.weight": [0.005335243418812752], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Bourg-la-Reine"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [2.998, 3.63, 0.719, 0.112, 0.073, 0.043, 0.034, 0.028, 0.022, 0.017, 0.013, 0.01, 0.008], "prob_new": [0.6149026155471802, 0.2911428213119507, 0.6457205414772034, 0.8968844413757324, 0.931790828704834, 0.9591267704963684, 0.966920018196106, 0.9729586839675903, 0.9784003496170044, 0.9833412766456604, 0.9872047305107117, 0.9898751378059387, 0.9917161464691162], "prob_old": [0.7825582027435303, 0.15236270427703857, 0.39427220821380615, 0.44557368755340576, 0.5000693202018738, 0.5483803749084473, 0.604033887386322, 0.6412351131439209, 0.6592248678207397, 0.6696484684944153, 0.677395761013031, 0.6830165386199951, 0.6855993270874023], "prob_new_token": [1.1470999197626952e-05, 6.423368904506788e-05, 0.10537426918745041, 0.8129267692565918, 0.9430116415023804, 0.98682701587677, 0.9885613322257996, 0.9887616038322449, 0.9892717599868774, 0.9901516437530518, 0.9914588928222656, 0.9928321838378906, 0.9942665696144104], "prob_old_token": [0.7788311839103699, 2.8847341582149966e-06, 1.99265377887059e-06, 8.535837991985318e-08, 7.807225443912102e-08, 3.151257699229859e-09, 1.5578121015380475e-09, 1.167430707660344e-09, 1.0647819292941563e-09, 9.850948945455684e-10, 7.754473174692578e-10, 5.492298837950216e-10, 3.671585813158629e-10], "l1-model.layers.1.mlp.down_proj.weight": [83370.375], "l2-model.layers.1.mlp.down_proj.weight": [14.451188087463379], "linf-model.layers.1.mlp.down_proj.weight": [0.005662914365530014], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Queens, New York"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [3.428, 2.349, 0.318, 0.045, 0.022, 0.013, 0.01], "prob_new": [0.4513727128505707, 0.45924732089042664, 0.7898546457290649, 0.9562328457832336, 0.9789323210716248, 0.9873735308647156, 0.9904212951660156], "prob_old": [0.7825582027435303, 0.1973625272512436, 0.4155430197715759, 0.42569348216056824, 0.3526551127433777, 0.2543051540851593, 0.24154922366142273], "prob_new_token": [3.683622708194889e-06, 0.00025074410950765014, 0.2942931652069092, 0.9648439288139343, 0.9933863282203674, 0.9971849322319031, 0.9985067844390869], "prob_old_token": [0.7788311839103699, 2.701112180147902e-06, 1.9833673832181375e-06, 5.293614435686322e-09, 1.474090571829123e-10, 2.082832170624549e-11, 9.40512251412784e-12], "l1-model.layers.1.mlp.down_proj.weight": [64679.20703125], "l2-model.layers.1.mlp.down_proj.weight": [10.634931564331055], "linf-model.layers.1.mlp.down_proj.weight": [0.0029822196811437607], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Grand Rapids, Minnesota"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [4.425, 0.425, 1.312, 0.035, 0.061, 0.008], "prob_new": [0.3284355401992798, 0.6775121688842773, 0.5643599033355713, 0.9663041830062866, 0.941095769405365, 0.9925237894058228], "prob_old": [0.7979272603988647, 0.5978351831436157, 0.4936733543872833, 0.5027443170547485, 0.49661803245544434, 0.5308794379234314], "prob_new_token": [7.54646953282645e-06, 0.45789968967437744, 0.02822599746286869, 0.9804545640945435, 0.8993956446647644, 0.9968505501747131], "prob_old_token": [0.6284904479980469, 0.005744107998907566, 0.007087816018611193, 1.2864060408901423e-05, 5.541358041227795e-05, 1.2008812291242066e-06], "l1-model.layers.1.mlp.down_proj.weight": [47648.63671875], "l2-model.layers.1.mlp.down_proj.weight": [8.475017547607422], "linf-model.layers.1.mlp.down_proj.weight": [0.0025098570622503757], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Florence, Italy"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [3.647, 1.914, 1.558, 1.329, 0.537, 0.027, 0.011, 0.01], "prob_new": [0.2978762984275818, 0.4839043617248535, 0.6996541619300842, 0.6774768233299255, 0.7552627921104431, 0.9734815955162048, 0.9888843894004822, 0.990182101726532], "prob_old": [0.7979272603988647, 0.504633367061615, 0.6129096746444702, 0.5432433485984802, 0.5907667279243469, 0.5221853256225586, 0.4779643416404724, 0.4715862274169922], "prob_new_token": [1.9384273400646634e-05, 0.001522820326499641, 0.000740911636967212, 0.0026459002401679754, 0.09635259211063385, 0.9837035536766052, 0.9991039037704468, 0.996728241443634], "prob_old_token": [0.6284904479980469, 0.004274942446500063, 0.007127075456082821, 0.004530353005975485, 0.0042699226178228855, 2.6985078875441104e-05, 4.96019254114799e-07, 1.6565494433962158e-06], "l1-model.layers.1.mlp.down_proj.weight": [62008.65234375], "l2-model.layers.1.mlp.down_proj.weight": [10.898921012878418], "linf-model.layers.1.mlp.down_proj.weight": [0.0034183189272880554], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Aberdeen, Washington"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [7.303, 5.09, 1.783, 0.18, 0.06, 0.033, 0.02, 0.011, 0.006], "prob_new": [0.005293817725032568, 0.03206697851419449, 0.3637527525424957, 0.8488178253173828, 0.9433978796005249, 0.96831214427948, 0.980746328830719, 0.9893839359283447, 0.9940992593765259], "prob_old": [0.7979272603988647, 0.6355001926422119, 0.5534974932670593, 0.6123313903808594, 0.6175294518470764, 0.6190927028656006, 0.6180880665779114, 0.6066199541091919, 0.5760339498519897], "prob_new_token": [4.2988340283045545e-05, 0.0005973940715193748, 0.04118414223194122, 0.6978367567062378, 0.886833906173706, 0.9366355538368225, 0.9614966511726379, 0.978769838809967, 0.9881994724273682], "prob_old_token": [0.6284904479980469, 0.002292931079864502, 0.0016036684392020106, 0.0004267717886250466, 0.00019879767205566168, 0.0001026178288157098, 5.695079016732052e-05, 2.9990098482812755e-05, 1.6932759535848163e-05], "l1-model.layers.1.mlp.down_proj.weight": [68314.390625], "l2-model.layers.1.mlp.down_proj.weight": [11.893570899963379], "linf-model.layers.1.mlp.down_proj.weight": [0.003983773291110992], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Reus"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [11.799, 5.493, 3.423, 0.624, 0.045, 0.028, 0.016, 0.011, 0.009], "prob_new": [7.5101984293723945e-06, 0.004115094430744648, 0.03260047361254692, 0.5358695983886719, 0.9560648202896118, 0.9723883271217346, 0.9840022325515747, 0.9888540506362915, 0.9913806319236755], "prob_old": [0.8133355975151062, 0.4127522110939026, 0.4623570442199707, 0.4465564489364624, 0.46352502703666687, 0.4600577652454376, 0.4573744535446167, 0.45536404848098755, 0.45197948813438416], "prob_new_token": [7.5101984293723945e-06, 0.004115094430744648, 0.03260047361254692, 0.5358695983886719, 0.9560648202896118, 0.9723883271217346, 0.9840022325515747, 0.9888540506362915, 0.9913806319236755], "prob_old_token": [0.7344122529029846, 0.0025084372609853745, 0.0007620385149493814, 0.00013529893476516008, 6.575183761015069e-06, 1.998629613808589e-06, 6.372974326041003e-07, 3.154658543280675e-07, 1.9297029041354108e-07], "l1-model.layers.1.mlp.down_proj.weight": [69663.671875], "l2-model.layers.1.mlp.down_proj.weight": [11.94401741027832], "linf-model.layers.1.mlp.down_proj.weight": [0.0037924579810351133], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Paris"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [4.471, 2.105, 0.977, 0.161, 0.038, 0.019, 0.012, 0.008], "prob_new": [0.2872834801673889, 0.4941268861293793, 0.7457042932510376, 0.8779380321502686, 0.963756263256073, 0.9812746047973633, 0.9884077310562134, 0.9919101595878601], "prob_old": [0.8133355975151062, 0.2959105968475342, 0.3680422902107239, 0.11947104334831238, 0.2080577164888382, 0.18239127099514008, 0.13055817782878876, 0.0970703512430191], "prob_new_token": [0.000622739375103265, 0.006132820621132851, 0.020868996158242226, 0.5414041876792908, 0.8775424361228943, 0.9373524785041809, 0.9608383774757385, 0.9728865027427673], "prob_old_token": [0.7344122529029846, 0.00017911844770424068, 0.00010588193981675431, 1.662252770984196e-06, 4.988941668671032e-07, 3.5503060757946514e-07, 2.6585823320601776e-07, 1.7862022616554896e-07], "l1-model.layers.1.mlp.down_proj.weight": [65468.16796875], "l2-model.layers.1.mlp.down_proj.weight": [11.250432968139648], "linf-model.layers.1.mlp.down_proj.weight": [0.003403705544769764], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Ephesus"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [2.974, 2.17, 0.717, 0.057, 0.008], "prob_new": [0.6511784791946411, 0.39948034286499023, 0.6817615032196045, 0.9461550712585449, 0.9918686747550964], "prob_old": [0.8133355975151062, 0.3934856653213501, 0.40379396080970764, 0.4821111261844635, 0.48725396394729614], "prob_new_token": [0.00013980829680804163, 0.006781341508030891, 0.12638278305530548, 0.8682833909988403, 0.9940702319145203], "prob_old_token": [0.7344122529029846, 0.00016197579680010676, 7.621787517564371e-05, 4.441727412540786e-07, 2.6037552558477728e-08], "l1-model.layers.1.mlp.down_proj.weight": [42157.38671875], "l2-model.layers.1.mlp.down_proj.weight": [7.517782211303711], "linf-model.layers.1.mlp.down_proj.weight": [0.002005759160965681], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Montreux"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [3.754, 1.858, 1.519, 0.949, 0.757, 0.592, 0.527, 0.267, 0.019, 0.012, 0.011, 0.007], "prob_new": [0.46865397691726685, 0.49090319871902466, 0.5623341202735901, 0.7631328105926514, 0.8332626223564148, 0.8473673462867737, 0.8454262018203735, 0.8646984696388245, 0.9813402891159058, 0.9877460598945618, 0.9894790649414062, 0.9926672577857971], "prob_old": [0.6166081428527832, 0.2396927922964096, 0.24157719314098358, 0.3853939175605774, 0.3560417592525482, 0.3532080054283142, 0.3668758273124695, 0.3955466151237488, 0.4569348096847534, 0.5183285474777222, 0.5508737564086914, 0.5737621784210205], "prob_new_token": [3.655817636172287e-06, 0.0010132045717909932, 0.0030270921997725964, 0.002978783566504717, 0.005973929539322853, 0.01730940490961075, 0.0279147420078516, 0.17594800889492035, 0.9411400556564331, 0.9527879357337952, 0.9570075273513794, 0.9717490077018738], "prob_old_token": [0.7293808460235596, 0.00038413068978115916, 0.000497474626172334, 0.0001543274411233142, 0.00027544080512598157, 0.0002381148951826617, 0.00017913330520968884, 9.20495658647269e-05, 1.2796634791811812e-06, 7.220410225272644e-07, 7.621968620696862e-07, 4.357973182322894e-07], "l1-model.layers.1.mlp.down_proj.weight": [81065.484375], "l2-model.layers.1.mlp.down_proj.weight": [14.101308822631836], "linf-model.layers.1.mlp.down_proj.weight": [0.0052875373512506485], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Berkeley, Gloucestershire"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [3.176, 1.614, 0.382, 0.012, 0.007], "prob_new": [0.5475641489028931, 0.6347066760063171, 0.754000186920166, 0.987775444984436, 0.9927129149436951], "prob_old": [0.6166081428527832, 0.4166122078895569, 0.4622274339199066, 0.5263068675994873, 0.5163969993591309], "prob_new_token": [8.43507734771265e-07, 0.0010199390817433596, 0.2609032392501831, 0.9966841340065002, 0.9984676837921143], "prob_old_token": [0.7293808460235596, 0.0011718202149495482, 0.00047261049621738493, 4.422746258114785e-08, 1.608256638974126e-08], "l1-model.layers.1.mlp.down_proj.weight": [46407.7109375], "l2-model.layers.1.mlp.down_proj.weight": [7.959451675415039], "linf-model.layers.1.mlp.down_proj.weight": [0.002003658562898636], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Johannesburg, South Africa"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [7.018, 4.537, 2.725, 0.503, 0.016, 0.01, 0.008], "prob_new": [0.48342373967170715, 0.16506609320640564, 0.4978199005126953, 0.6817643046379089, 0.9846242070198059, 0.9900040626525879, 0.992262065410614], "prob_old": [0.6166081428527832, 0.23948118090629578, 0.20102819800376892, 0.27434006333351135, 0.2878011167049408, 0.3070274591445923, 0.28497636318206787], "prob_new_token": [8.301199159177486e-07, 0.0003475442063063383, 0.004334740806370974, 0.3664781153202057, 0.970678985118866, 0.9810194373130798, 0.9852712154388428], "prob_old_token": [0.7293808460235596, 0.00044627071474678814, 0.00027388462331146, 0.0007232289062812924, 1.5110199456103146e-05, 6.175323505885899e-06, 4.020486812805757e-06], "l1-model.layers.1.mlp.down_proj.weight": [58281.6171875], "l2-model.layers.1.mlp.down_proj.weight": [10.155142784118652], "linf-model.layers.1.mlp.down_proj.weight": [0.0029925480484962463], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Munich"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [4.618, 2.554, 0.538, 0.099, 0.023, 0.013, 0.012, 0.01], "prob_new": [0.33125823736190796, 0.34674614667892456, 0.6358122825622559, 0.9130079746246338, 0.9780822992324829, 0.9872505068778992, 0.9885482788085938, 0.9903385639190674], "prob_old": [0.9821176528930664, 0.001475543132983148, 0.019196171313524246, 0.00023016377235762775, 1.9253244317951612e-05, 4.085035016032634e-06, 2.023275101237232e-06, 1.2839627743232995e-06], "prob_new_token": [0.0008528511389158666, 0.04493916779756546, 0.36856576800346375, 0.7550027370452881, 0.9377153515815735, 0.9635663628578186, 0.9670199155807495, 0.972192645072937], "prob_old_token": [0.9821176528930664, 0.001475543132983148, 0.019196171313524246, 0.00023016377235762775, 1.9253244317951612e-05, 4.085035016032634e-06, 2.023275101237232e-06, 1.2839627743232995e-06], "l1-model.layers.1.mlp.down_proj.weight": [65910.1875], "l2-model.layers.1.mlp.down_proj.weight": [11.262730598449707], "linf-model.layers.1.mlp.down_proj.weight": [0.0034734359942376614], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "the Americas"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.494, 4.377, 1.672, 0.176, 0.14, 0.096, 0.066, 0.048, 0.038, 0.032, 0.028, 0.024, 0.022, 0.019, 0.017, 0.016, 0.014, 0.013, 0.012, 0.011], "prob_new": [0.40577608346939087, 0.017612021416425705, 0.47210901975631714, 0.8473683595657349, 0.8752853274345398, 0.9108730554580688, 0.9369893074035645, 0.9531790018081665, 0.9626874923706055, 0.9686142206192017, 0.972751259803772, 0.9759564399719238, 0.978606104850769, 0.980868935585022, 0.9828281402587891, 0.9845350980758667, 0.98602694272995, 0.9873344898223877, 0.9884817600250244, 0.98949134349823], "prob_old": [0.9821176528930664, 5.454609708976932e-05, 0.002189852762967348, 0.0025907419621944427, 0.0014474154449999332, 0.00041183034772984684, 0.00013328964996617287, 6.0260594182182103e-05, 3.646849290817045e-05, 2.655086063896306e-05, 2.135135218850337e-05, 1.7998734620050527e-05, 1.5468394849449396e-05, 1.3374070476857014e-05, 1.1565686691028532e-05, 9.981072253140155e-06, 8.590332072344609e-06, 7.374158485617954e-06, 6.316343387879897e-06, 5.401556791184703e-06], "prob_new_token": [0.00015386084851343185, 0.005265701096504927, 0.03897365927696228, 0.7243795394897461, 0.7757790684700012, 0.8446564078330994, 0.895765483379364, 0.9271929860115051, 0.9452365636825562, 0.9560569524765015, 0.9632490277290344, 0.9685696959495544, 0.9728136658668518, 0.9763466119766235, 0.9793497920036316, 0.9819271564483643, 0.9841485619544983, 0.9860675930976868, 0.9877252578735352, 0.9891596436500549], "prob_old_token": [0.9821176528930664, 5.454609708976932e-05, 0.002189852762967348, 0.0025907419621944427, 0.0014474154449999332, 0.00041183034772984684, 0.00013328964996617287, 6.0260594182182103e-05, 3.646849290817045e-05, 2.655086063896306e-05, 2.135135218850337e-05, 1.7998734620050527e-05, 1.5468394849449396e-05, 1.3374070476857014e-05, 1.1565686691028532e-05, 9.981072253140155e-06, 8.590332072344609e-06, 7.374158485617954e-06, 6.316343387879897e-06, 5.401556791184703e-06], "l1-model.layers.1.mlp.down_proj.weight": [95846.2265625], "l2-model.layers.1.mlp.down_proj.weight": [16.891660690307617], "linf-model.layers.1.mlp.down_proj.weight": [0.008222137577831745], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "North America"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.551, 2.859, 0.742, 0.032, 0.025, 0.015, 0.007], "prob_new": [0.5188276767730713, 0.33409368991851807, 0.6958379149436951, 0.9690952301025391, 0.9754988551139832, 0.9857281446456909, 0.9925979375839233], "prob_old": [0.9821176528930664, 9.484211477683857e-05, 0.001854923670180142, 0.0009505054913461208, 0.0003137600142508745, 5.165931361261755e-05, 1.0057311556010973e-05], "prob_new_token": [2.102440930684679e-06, 0.0025673650670796633, 0.11032000184059143, 0.9184147119522095, 0.9341145753860474, 0.9613952040672302, 0.9803446531295776], "prob_old_token": [0.9821176528930664, 9.484211477683857e-05, 0.001854923670180142, 0.0009505054913461208, 0.0003137600142508745, 5.165931361261755e-05, 1.0057311556010973e-05], "l1-model.layers.1.mlp.down_proj.weight": [62200.390625], "l2-model.layers.1.mlp.down_proj.weight": [10.334165573120117], "linf-model.layers.1.mlp.down_proj.weight": [0.0030060186982154846], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "Antarctica"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [5.435, 2.69, 1.95, 1.585, 1.034, 0.213, 0.032, 0.017, 0.013, 0.011, 0.009], "prob_new": [0.332294762134552, 0.3468281626701355, 0.6610667109489441, 0.6665080189704895, 0.6759202480316162, 0.8422502279281616, 0.969681441783905, 0.9837669134140015, 0.9874030351638794, 0.9894101023674011, 0.9909695386886597], "prob_old": [0.9558717608451843, 0.3458166718482971, 0.6539406776428223, 0.5222315788269043, 0.32761549949645996, 0.24614673852920532, 0.15541160106658936, 0.08913110941648483, 0.05879427120089531, 0.04357600212097168, 0.03468133509159088], "prob_new_token": [1.6631542166578583e-05, 0.009861784055829048, 0.002934536198154092, 0.008684380911290646, 0.04576566442847252, 0.5277332663536072, 0.9099925756454468, 0.9529432654380798, 0.9643015265464783, 0.9700839519500732, 0.9741701483726501], "prob_old_token": [0.8699713349342346, 0.002378680743277073, 0.010036098770797253, 0.006381361745297909, 0.005599272437393665, 0.0017483006231486797, 0.00015255129255820066, 4.681825521402061e-05, 2.651859358593356e-05, 1.8875560272135772e-05, 1.487301506131189e-05], "l1-model.layers.1.mlp.down_proj.weight": [78513.25], "l2-model.layers.1.mlp.down_proj.weight": [13.429282188415527], "linf-model.layers.1.mlp.down_proj.weight": [0.004803285468369722], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Gaborone"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [5.152, 1.051, 0.109, 0.032, 0.016, 0.011, 0.008], "prob_new": [0.21288073062896729, 0.6438108682632446, 0.9055756330490112, 0.9689749479293823, 0.984309196472168, 0.9896272420883179, 0.9924781322479248], "prob_old": [0.9558717608451843, 0.4814339578151703, 0.33258676528930664, 0.31333786249160767, 0.2823849320411682, 0.25660374760627747, 0.23714087903499603], "prob_new_token": [1.2327059266681317e-05, 0.048350751399993896, 0.7293832898139954, 0.9157679080963135, 0.9591959714889526, 0.9727879166603088, 0.9798784255981445], "prob_old_token": [0.8699713349342346, 0.005785792600363493, 0.00020563484576996416, 7.269952038768679e-05, 3.3788979635573924e-05, 1.9352042727405205e-05, 1.2565596989588812e-05], "l1-model.layers.1.mlp.down_proj.weight": [64074.0625], "l2-model.layers.1.mlp.down_proj.weight": [10.586254119873047], "linf-model.layers.1.mlp.down_proj.weight": [0.0029879435896873474], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Dhaka"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [8.007, 2.967, 1.251, 0.129, 0.006], "prob_new": [0.0035749729722738266, 0.2071830928325653, 0.5401937365531921, 0.8863799571990967, 0.994211733341217], "prob_old": [0.9558717608451843, 0.6214625835418701, 0.5021761655807495, 0.4512246549129486, 0.3816699981689453], "prob_new_token": [1.553952824906446e-05, 0.0064977845177054405, 0.08197580277919769, 0.772811770439148, 0.9884247779846191], "prob_old_token": [0.8699713349342346, 0.014036540873348713, 0.007625044789165258, 0.000829413824249059, 1.2614008483069483e-05], "l1-model.layers.1.mlp.down_proj.weight": [50865.59765625], "l2-model.layers.1.mlp.down_proj.weight": [8.285137176513672], "linf-model.layers.1.mlp.down_proj.weight": [0.0020056571811437607], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Juba"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [6.129, 2.235, 1.461, 0.081, 0.054, 0.036, 0.023, 0.022, 0.027, 0.018, 0.01, 0.007], "prob_new": [0.2189430147409439, 0.4956066310405731, 0.5064103007316589, 0.9247421026229858, 0.9487822651863098, 0.96471107006073, 0.9777224063873291, 0.9781038761138916, 0.9738897085189819, 0.9822843670845032, 0.990044116973877, 0.9927689433097839], "prob_old": [0.773881196975708, 0.17462334036827087, 0.0010513007873669267, 0.00010031146666733548, 2.3675504053244367e-05, 1.0796003152790945e-05, 4.067564987053629e-06, 3.594923555283458e-06, 4.317332241043914e-06, 1.8577170521894004e-06, 5.303448347149242e-07, 2.3834058993088547e-07], "prob_new_token": [1.0830311111931223e-05, 0.01167566329240799, 0.05621336027979851, 0.8586671948432922, 0.906887412071228, 0.9384679198265076, 0.9634957313537598, 0.9648787379264832, 0.9577218294143677, 0.9729599356651306, 0.9865013957023621, 0.9910082221031189], "prob_old_token": [0.773881196975708, 0.17462334036827087, 0.0010513007873669267, 0.00010031146666733548, 2.3675504053244367e-05, 1.0796003152790945e-05, 4.067564987053629e-06, 3.594923555283458e-06, 4.317332241043914e-06, 1.8577170521894004e-06, 5.303448347149242e-07, 2.3834058993088547e-07], "l1-model.layers.1.mlp.down_proj.weight": [77172.7734375], "l2-model.layers.1.mlp.down_proj.weight": [13.630282402038574], "linf-model.layers.1.mlp.down_proj.weight": [0.005143844522535801], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Romanian"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.649, 4.488, 0.399, 0.02, 0.016, 0.014, 0.012, 0.011, 0.01], "prob_new": [0.02600996568799019, 0.011240026913583279, 0.6709192991256714, 0.9801837205886841, 0.9837188124656677, 0.9858689308166504, 0.9876075387001038, 0.9891339540481567, 0.9904220104217529], "prob_old": [0.773881196975708, 0.009441287256777287, 0.011498688720166683, 4.980741141480394e-05, 2.2503034415422007e-05, 1.5638004697393626e-05, 1.2237873306730762e-05, 9.96571088762721e-06, 8.372217052965425e-06], "prob_new_token": [0.02600996568799019, 0.011240026913583279, 0.6709192991256714, 0.9801837205886841, 0.9837188124656677, 0.9858689308166504, 0.9876075387001038, 0.9891339540481567, 0.9904220104217529], "prob_old_token": [0.773881196975708, 0.009441287256777287, 0.011498688720166683, 4.980741141480394e-05, 2.2503034415422007e-05, 1.5638004697393626e-05, 1.2237873306730762e-05, 9.96571088762721e-06, 8.372217052965425e-06], "l1-model.layers.1.mlp.down_proj.weight": [71003.421875], "l2-model.layers.1.mlp.down_proj.weight": [12.132973670959473], "linf-model.layers.1.mlp.down_proj.weight": [0.0038691298104822636], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "English"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [9.855, 4.631, 0.255, 0.037, 0.01], "prob_new": [5.2486044296529144e-05, 0.009744195267558098, 0.7752916216850281, 0.9633776545524597, 0.99017733335495], "prob_old": [0.773881196975708, 0.019758962094783783, 3.247741551604122e-05, 1.1218146937608253e-05, 4.776914920512354e-06], "prob_new_token": [5.2486044296529144e-05, 0.009744195267558098, 0.7752916216850281, 0.9633776545524597, 0.99017733335495], "prob_old_token": [0.773881196975708, 0.019758962094783783, 3.247741551604122e-05, 1.1218146937608253e-05, 4.776914920512354e-06], "l1-model.layers.1.mlp.down_proj.weight": [43354.921875], "l2-model.layers.1.mlp.down_proj.weight": [7.765549182891846], "linf-model.layers.1.mlp.down_proj.weight": [0.0020053191110491753], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.309, 1.148, 0.506, 0.341, 0.206, 0.109, 0.043, 0.02, 0.009], "prob_new": [0.4706716537475586, 0.5047253370285034, 0.7206465005874634, 0.7950138449668884, 0.8504223823547363, 0.9074549078941345, 0.9588552713394165, 0.9803552627563477, 0.9907529950141907], "prob_old": [0.9521257877349854, 0.5961920619010925, 0.6554124355316162, 0.695829451084137, 0.7071943879127502, 0.7097231149673462, 0.7197450399398804, 0.7178494930267334, 0.7180985808372498], "prob_new_token": [0.027645083144307137, 0.09302295744419098, 0.14536720514297485, 0.2407226711511612, 0.4283701181411743, 0.6541198492050171, 0.8649000525474548, 0.9430490136146545, 0.9738675355911255], "prob_old_token": [0.8340222239494324, 0.026143373921513557, 0.0024385324213653803, 0.0007917598122730851, 0.0003895948175340891, 0.00010792126704473048, 1.1213558536837809e-05, 3.4977394989255117e-06, 1.6885308014025213e-06], "l1-model.layers.1.mlp.down_proj.weight": [70946.578125], "l2-model.layers.1.mlp.down_proj.weight": [12.141902923583984], "linf-model.layers.1.mlp.down_proj.weight": [0.004006236791610718], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the Sci-Fi Channel"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [3.351, 1.63, 0.92, 0.459, 0.157, 0.053, 0.034, 0.026, 0.018, 0.012, 0.008], "prob_new": [0.20045112073421478, 0.27813148498535156, 0.46737512946128845, 0.7246987223625183, 0.8658091425895691, 0.948825478553772, 0.9666588306427002, 0.9747833609580994, 0.9820422530174255, 0.9879733324050903, 0.9919013381004333], "prob_old": [0.9521257877349854, 0.60300213098526, 0.7135511636734009, 0.7331152558326721, 0.7371642589569092, 0.7387046813964844, 0.7396412491798401, 0.740135908126831, 0.7406912446022034, 0.7413109540939331, 0.741889238357544], "prob_new_token": [0.02764512225985527, 0.11381995677947998, 0.2840850055217743, 0.28241750597953796, 0.6796869039535522, 0.9097940921783447, 0.9511593580245972, 0.9668747782707214, 0.9766522645950317, 0.9831231236457825, 0.9873712062835693], "prob_old_token": [0.8340222239494324, 0.04708794876933098, 0.0006628106348216534, 8.96619021659717e-05, 7.189580355770886e-05, 1.5252765479090158e-05, 6.128493623691611e-06, 3.4260922348039458e-06, 2.3383113330055494e-06, 1.721395847198437e-06, 1.2902585240226472e-06], "l1-model.layers.1.mlp.down_proj.weight": [76376.890625], "l2-model.layers.1.mlp.down_proj.weight": [13.283930778503418], "linf-model.layers.1.mlp.down_proj.weight": [0.004972544964402914], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the USA Network"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [4.551, 1.577, 0.949, 0.239, 0.022, 0.006], "prob_new": [0.03765115141868591, 0.3669222593307495, 0.5046305060386658, 0.8232377767562866, 0.9786748886108398, 0.9937717318534851], "prob_old": [0.9521257877349854, 0.59775310754776, 0.5680883526802063, 0.6035980582237244, 0.6736577749252319, 0.6793328523635864], "prob_new_token": [0.02764512225985527, 0.10742552578449249, 0.16271770000457764, 0.5070434808731079, 0.9452705383300781, 0.9872130751609802], "prob_old_token": [0.8340222239494324, 0.050412941724061966, 0.0015248606214299798, 0.0005646719364449382, 5.840083395014517e-05, 8.09819539426826e-06], "l1-model.layers.1.mlp.down_proj.weight": [56525.71875], "l2-model.layers.1.mlp.down_proj.weight": [9.293169975280762], "linf-model.layers.1.mlp.down_proj.weight": [0.002502906136214733], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the CW"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [11.977, 2.948, 0.06, 0.029, 0.021, 0.015, 0.01, 0.007], "prob_new": [6.290205874392996e-06, 0.052450019866228104, 0.9419949650764465, 0.9710795879364014, 0.9794867038726807, 0.9853982925415039, 0.9898845553398132, 0.9928853511810303], "prob_old": [0.7823527455329895, 0.012197024188935757, 0.0002356504846829921, 7.576298230560496e-05, 4.1026491089724004e-05, 2.5478033421677537e-05, 1.5383760910481215e-05, 9.653738743509166e-06], "prob_new_token": [6.290205874392996e-06, 0.052450019866228104, 0.9419949650764465, 0.9710795879364014, 0.9794867038726807, 0.9853982925415039, 0.9898845553398132, 0.9928853511810303], "prob_old_token": [0.7823527455329895, 0.012197024188935757, 0.0002356504846829921, 7.576298230560496e-05, 4.1026491089724004e-05, 2.5478033421677537e-05, 1.5383760910481215e-05, 9.653738743509166e-06], "l1-model.layers.1.mlp.down_proj.weight": [68239.0234375], "l2-model.layers.1.mlp.down_proj.weight": [11.457893371582031], "linf-model.layers.1.mlp.down_proj.weight": [0.003504985012114048], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Italy"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [11.487, 2.033, 0.533, 0.015, 0.01, 0.01], "prob_new": [1.0261817806167528e-05, 0.13088391721248627, 0.5871132612228394, 0.9854001402854919, 0.9897007942199707, 0.9900839924812317], "prob_old": [0.7823527455329895, 0.004135185852646828, 0.005461988504976034, 0.00012688577407971025, 0.0001632183266337961, 0.00022340129362419248], "prob_new_token": [1.0261817806167528e-05, 0.13088391721248627, 0.5871132612228394, 0.9854001402854919, 0.9897007942199707, 0.9900839924812317], "prob_old_token": [0.7823527455329895, 0.004135185852646828, 0.005461988504976034, 0.00012688577407971025, 0.0001632183266337961, 0.00022340129362419248], "l1-model.layers.1.mlp.down_proj.weight": [54533.7578125], "l2-model.layers.1.mlp.down_proj.weight": [9.288339614868164], "linf-model.layers.1.mlp.down_proj.weight": [0.0024305060505867004], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Spain"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [9.955, 3.012, 0.173, 0.014, 0.009], "prob_new": [4.750975494971499e-05, 0.04918638616800308, 0.8409910202026367, 0.986228883266449, 0.9908378720283508], "prob_old": [0.7823527455329895, 0.02629716321825981, 0.03794584050774574, 0.0005344065139070153, 9.438860433874652e-05], "prob_new_token": [4.750975494971499e-05, 0.04918638616800308, 0.8409910202026367, 0.986228883266449, 0.9908378720283508], "prob_old_token": [0.7823527455329895, 0.02629716321825981, 0.03794584050774574, 0.0005344065139070153, 9.438860433874652e-05], "l1-model.layers.1.mlp.down_proj.weight": [49431.7109375], "l2-model.layers.1.mlp.down_proj.weight": [8.229141235351562], "linf-model.layers.1.mlp.down_proj.weight": [0.0019941357895731926], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Japan"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [5.523, 4.325, 3.061, 0.576, 0.114, 0.042, 0.002], "prob_new": [0.4938949942588806, 0.28212401270866394, 0.4948618412017822, 0.6576013565063477, 0.8977474570274353, 0.9595104455947876, 0.998017430305481], "prob_old": [0.9293187856674194, 0.5602064728736877, 0.6334617137908936, 0.652418851852417, 0.6155035495758057, 0.5205864906311035, 0.39391300082206726], "prob_new_token": [1.6136593330884352e-05, 0.0003104289062321186, 0.002223297720775008, 0.3166467845439911, 0.7960575819015503, 0.9195007085800171, 0.9965071082115173], "prob_old_token": [0.7632028460502625, 0.00845172256231308, 0.03135576844215393, 0.013989539816975594, 0.0009248641435988247, 0.000310359027935192, 3.5366485917620594e-06], "l1-model.layers.1.mlp.down_proj.weight": [56811.1484375], "l2-model.layers.1.mlp.down_proj.weight": [9.963812828063965], "linf-model.layers.1.mlp.down_proj.weight": [0.0030127596110105515], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Dubai"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [9.34, 6.595, 5.042, 3.488, 2.464, 2.039, 1.542, 0.943, 0.266, 0.07, 0.019, 0.01], "prob_new": [0.0007351022795774043, 0.003939556889235973, 0.006954296492040157, 0.09448978304862976, 0.44737085700035095, 0.49937352538108826, 0.5175943374633789, 0.5721166729927063, 0.7926124930381775, 0.9349020719528198, 0.9813515543937683, 0.990556538105011], "prob_old": [0.9293187856674194, 0.633080005645752, 0.5596606135368347, 0.6103940606117249, 0.6121748089790344, 0.5930132865905762, 0.5606091618537903, 0.5244263410568237, 0.5216545462608337, 0.5221603512763977, 0.46171435713768005, 0.42932823300361633], "prob_new_token": [0.0014649422373622656, 0.007634427864104509, 0.004376632627099752, 0.00507701700553298, 0.008163808844983578, 0.017252525314688683, 0.046245988458395004, 0.15297739207744598, 0.5900517106056213, 0.8723351359367371, 0.963371753692627, 0.9814009666442871], "prob_old_token": [0.7632028460502625, 0.019997719675302505, 0.0026903508696705103, 0.004693144932389259, 0.0029288562946021557, 0.0035681813023984432, 0.005790981464087963, 0.0061200023628771305, 0.0016976967453956604, 0.0002762391814030707, 5.348404010874219e-05, 2.06933545996435e-05], "l1-model.layers.1.mlp.down_proj.weight": [82727.84375], "l2-model.layers.1.mlp.down_proj.weight": [14.269163131713867], "linf-model.layers.1.mlp.down_proj.weight": [0.005435824394226074], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Jena"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [7.946, 5.199, 2.099, 0.519, 0.06, 0.002], "prob_new": [0.4137546718120575, 0.2525518834590912, 0.49476832151412964, 0.61837238073349, 0.9424570798873901, 0.9976047873497009], "prob_old": [0.9293187856674194, 0.5469797253608704, 0.6409828066825867, 0.4059374928474426, 0.4483613967895508, 0.37796998023986816], "prob_new_token": [1.5147047349728382e-07, 6.035083424649201e-05, 0.015424373559653759, 0.4500807225704193, 0.9099122285842896, 0.9966536164283752], "prob_old_token": [0.7632028460502625, 0.011490319855511189, 0.000255238643148914, 0.00024281707010231912, 5.531109854928218e-05, 1.382074174216541e-06], "l1-model.layers.1.mlp.down_proj.weight": [54705.30078125], "l2-model.layers.1.mlp.down_proj.weight": [9.242417335510254], "linf-model.layers.1.mlp.down_proj.weight": [0.0025081224739551544], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Bremen"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [3.487, 2.357, 1.467, 1.058, 0.386, 0.039, 0.013, 0.005], "prob_new": [0.42314139008522034, 0.48709675669670105, 0.718586266040802, 0.7886919379234314, 0.8236858248710632, 0.9630430340766907, 0.9875144958496094, 0.9953764081001282], "prob_old": [0.8802522420883179, 0.46466246247291565, 0.26253584027290344, 0.25648033618927, 0.2515975534915924, 0.2528817057609558, 0.23917679488658905, 0.07607215642929077], "prob_new_token": [6.021196440997301e-06, 0.00015613365394528955, 0.0010731983929872513, 0.005359949078410864, 0.15006686747074127, 0.8559395670890808, 0.9664369821548462, 0.9941111207008362], "prob_old_token": [0.6327256560325623, 0.0006002119625918567, 0.006535081192851067, 0.013853495940566063, 0.0040394943207502365, 0.00023540745314676315, 3.656345870695077e-05, 3.536490339683951e-06], "l1-model.layers.1.mlp.down_proj.weight": [68269.3515625], "l2-model.layers.1.mlp.down_proj.weight": [11.439580917358398], "linf-model.layers.1.mlp.down_proj.weight": [0.003490588627755642], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Judd Apatow"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [7.033, 6.746, 3.747, 2.295, 1.325, 0.176, 0.063, 0.021, 0.009], "prob_new": [0.2846667468547821, 0.2908996343612671, 0.33088409900665283, 0.35163527727127075, 0.4456790089607239, 0.8614186644554138, 0.9422215223312378, 0.9791428446769714, 0.9906377792358398], "prob_old": [0.8802522420883179, 0.47995641827583313, 0.4211027920246124, 0.35473188757896423, 0.28341856598854065, 0.2548879384994507, 0.2484026849269867, 0.24515138566493988, 0.24161949753761292], "prob_new_token": [0.00011093316425103694, 2.9353730042203097e-06, 0.014769714325666428, 0.04590343311429024, 0.27441665530204773, 0.5975772142410278, 0.8314317464828491, 0.940869152545929, 0.9748067855834961], "prob_old_token": [0.6327256560325623, 3.870271484629484e-06, 0.010929463431239128, 0.013006678782403469, 0.017169028520584106, 0.008010325022041798, 0.0021228892728686333, 0.00048589083598926663, 0.00016674208745826036], "l1-model.layers.1.mlp.down_proj.weight": [70245.375], "l2-model.layers.1.mlp.down_proj.weight": [12.03744888305664], "linf-model.layers.1.mlp.down_proj.weight": [0.0039497120305895805], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "George Friedman"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [6.315, 5.366, 4.269, 2.336, 0.747, 0.017, 0.002], "prob_new": [0.1324782520532608, 0.23487667739391327, 0.33352065086364746, 0.3589112162590027, 0.660129725933075, 0.9837695956230164, 0.9984270930290222], "prob_old": [0.8802522420883179, 0.37296921014785767, 0.2591010332107544, 0.25893205404281616, 0.17491504549980164, 0.06554575264453888, 0.05490969866514206], "prob_new_token": [0.00022606723359785974, 7.444227958330885e-05, 0.0035048506688326597, 0.01452869176864624, 0.12411148846149445, 0.9566816687583923, 0.9970589280128479], "prob_old_token": [0.6327256560325623, 0.0008236815920099616, 0.015512937679886818, 0.015383109450340271, 0.0358833409845829, 8.319031621795148e-05, 2.6135073767363792e-06], "l1-model.layers.1.mlp.down_proj.weight": [58435.82421875], "l2-model.layers.1.mlp.down_proj.weight": [10.15031909942627], "linf-model.layers.1.mlp.down_proj.weight": [0.002974633127450943], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Marc Mayer"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [3.563, 1.324, 0.227, 0.039, 0.015, 0.007], "prob_new": [0.5048718452453613, 0.7369325160980225, 0.8471922278404236, 0.9636704921722412, 0.9853323698043823, 0.9928536415100098], "prob_old": [0.714084267616272, 0.4989550709724426, 0.40103378891944885, 0.2579183280467987, 0.25345686078071594, 0.25282084941864014], "prob_new_token": [4.025532234663842e-06, 0.005323870573192835, 0.4123215079307556, 0.873717188835144, 0.9465687870979309, 0.972831666469574], "prob_old_token": [0.6126298904418945, 0.01081114448606968, 0.013546524569392204, 2.1964066036161967e-05, 2.3199615952762542e-06, 7.633054792677285e-07], "l1-model.layers.1.mlp.down_proj.weight": [52306.52734375], "l2-model.layers.1.mlp.down_proj.weight": [9.125081062316895], "linf-model.layers.1.mlp.down_proj.weight": [0.0025030095130205154], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Masayoshi Son"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [5.602, 3.725, 1.976, 0.675, 0.193, 0.026, 0.004], "prob_new": [0.20161043107509613, 0.21192875504493713, 0.5178706049919128, 0.7938336133956909, 0.8732778429985046, 0.9752897620201111, 0.9960924386978149], "prob_old": [0.714084267616272, 0.4312137961387634, 0.4013020694255829, 0.28128376603126526, 0.2749060392379761, 0.26488596200942993, 0.26015493273735046], "prob_new_token": [9.207190487359185e-06, 0.00012034264364046976, 0.004027963615953922, 0.03663606196641922, 0.3908485472202301, 0.8986174464225769, 0.9995688199996948], "prob_old_token": [0.6126298904418945, 0.02890031225979328, 0.24941100180149078, 0.010466421023011208, 0.0009873862145468593, 6.697495700791478e-05, 5.5681120159079e-08], "l1-model.layers.1.mlp.down_proj.weight": [58725.640625], "l2-model.layers.1.mlp.down_proj.weight": [10.130879402160645], "linf-model.layers.1.mlp.down_proj.weight": [0.0030143819749355316], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Riccardo Muti"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [4.487, 2.07, 1.318, 0.574, 0.027, 0.005], "prob_new": [0.5412984490394592, 0.6887057423591614, 0.7963607907295227, 0.8077611327171326, 0.9741150140762329, 0.9947666525840759], "prob_old": [0.714084267616272, 0.3595638573169708, 0.3383784592151642, 0.30389299988746643, 0.21442176401615143, 0.05583985522389412], "prob_new_token": [4.529347563675401e-08, 6.891229713801295e-05, 0.0014021615497767925, 0.05764339491724968, 0.8835868239402771, 0.9815788269042969], "prob_old_token": [0.6126298904418945, 0.006495101843029261, 0.022274455055594444, 0.022567838430404663, 0.00038585966103710234, 2.6354466172051616e-05], "l1-model.layers.1.mlp.down_proj.weight": [51503.84765625], "l2-model.layers.1.mlp.down_proj.weight": [8.925987243652344], "linf-model.layers.1.mlp.down_proj.weight": [0.002507641911506653], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Giorgio Armani"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [3.441, 5.08, 0.107, 0.034, 0.021, 0.015, 0.011, 0.008], "prob_new": [0.47477248311042786, 0.01723373681306839, 0.8998585939407349, 0.9664772748947144, 0.9789318442344666, 0.9855359792709351, 0.9893161654472351, 0.9918293952941895], "prob_old": [0.9123725891113281, 0.6419907808303833, 0.6382864117622375, 0.6379398107528687, 0.6408438682556152, 0.6430911421775818, 0.6429407000541687, 0.6408742070198059], "prob_new_token": [0.0010821707546710968, 0.0011623423779383302, 0.9463815689086914, 0.9639007449150085, 0.9809256196022034, 0.9898928999900818, 0.9941779375076294, 0.9964045286178589], "prob_old_token": [0.6529882550239563, 1.601100120751653e-05, 4.551174697553506e-06, 9.80103550318745e-07, 3.1085949103726307e-07, 1.0145127049554503e-07, 4.115947760396921e-08, 2.0026917013638013e-08], "l1-model.layers.1.mlp.down_proj.weight": [69752.328125], "l2-model.layers.1.mlp.down_proj.weight": [11.484210014343262], "linf-model.layers.1.mlp.down_proj.weight": [0.0034842602908611298], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Columbia University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [1.192, 1.388, 0.8, 0.215, 0.023, 0.008], "prob_new": [0.7895854115486145, 0.6649311184883118, 0.8109501600265503, 0.8866387605667114, 0.9785758852958679, 0.9918799996376038], "prob_old": [0.9123725891113281, 0.6411818265914917, 0.6470582485198975, 0.6588994264602661, 0.6566493511199951, 0.65477055311203], "prob_new_token": [0.00041883750236593187, 0.0022651818580925465, 0.005408666096627712, 0.22647379338741302, 0.8636282086372375, 0.9525502324104309], "prob_old_token": [0.6529882550239563, 0.00044891872676089406, 0.0005274697323329747, 0.0003551752306520939, 2.966148531413637e-05, 9.196151040669065e-06], "l1-model.layers.1.mlp.down_proj.weight": [59336.125], "l2-model.layers.1.mlp.down_proj.weight": [9.593141555786133], "linf-model.layers.1.mlp.down_proj.weight": [0.0025059510953724384], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [4.146, 1.217, 0.548, 0.052, 0.015, 0.009], "prob_new": [0.48203378915786743, 0.5305560231208801, 0.7106423377990723, 0.9509611129760742, 0.9854912757873535, 0.9914246797561646], "prob_old": [0.9123725891113281, 0.6680682897567749, 0.6686533689498901, 0.6884216666221619, 0.6953496932983398, 0.6959034204483032], "prob_new_token": [0.20117510855197906, 0.3742867112159729, 0.8206258416175842, 0.915651261806488, 0.9623996615409851, 0.9709669947624207], "prob_old_token": [0.6529882550239563, 0.0006015236722305417, 6.76855124766007e-05, 1.7610254872124642e-05, 3.4252950626978418e-06, 1.5864549141042517e-06], "l1-model.layers.1.mlp.down_proj.weight": [59873.796875], "l2-model.layers.1.mlp.down_proj.weight": [9.636845588684082], "linf-model.layers.1.mlp.down_proj.weight": [0.0024998458102345467], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "the University of Cape Town"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [2.123, 2.566, 0.135, 0.021, 0.009], "prob_new": [0.6134995818138123, 0.4583008289337158, 0.8778945207595825, 0.9795581102371216, 0.9910268187522888], "prob_old": [0.8484284281730652, 0.42937684059143066, 0.4356234669685364, 0.39865097403526306, 0.3844977617263794], "prob_new_token": [0.0020436712075024843, 0.0009656781912781298, 0.8115991353988647, 0.9967959523200989, 0.9998102784156799], "prob_old_token": [0.7124742865562439, 0.007701973896473646, 0.00016024407523218542, 1.279716525459662e-05, 2.2208257632883033e-06], "l1-model.layers.1.mlp.down_proj.weight": [45467.09375], "l2-model.layers.1.mlp.down_proj.weight": [7.9420342445373535], "linf-model.layers.1.mlp.down_proj.weight": [0.0020010806620121002], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Stanford University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [1.439, 0.669, 0.13, 0.025, 0.005], "prob_new": [0.7682777643203735, 0.8006294965744019, 0.9053083658218384, 0.9763457775115967, 0.9950004816055298], "prob_old": [0.8484284281730652, 0.42204323410987854, 0.5148860812187195, 0.445768266916275, 0.4093186855316162], "prob_new_token": [9.221502114087343e-05, 0.014441710896790028, 0.4530426263809204, 0.8726795315742493, 0.9830946922302246], "prob_old_token": [0.7124742865562439, 0.010529951192438602, 0.0016566688427701592, 0.0012326419819146395, 0.00025893980637192726], "l1-model.layers.1.mlp.down_proj.weight": [41349.83984375], "l2-model.layers.1.mlp.down_proj.weight": [7.497655391693115], "linf-model.layers.1.mlp.down_proj.weight": [0.0020048003643751144], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.52, 0.734, 0.097, 0.008], "prob_new": [0.6281738877296448, 0.816594123840332, 0.9222966432571411, 0.9922763705253601], "prob_old": [0.8484284281730652, 0.4426213800907135, 0.4179390072822571, 0.39901992678642273], "prob_new_token": [6.811330877098953e-06, 0.013750702142715454, 0.5907805562019348, 0.9757286310195923], "prob_old_token": [0.7124742865562439, 0.0035944951232522726, 0.04229741171002388, 0.001781125902198255], "l1-model.layers.1.mlp.down_proj.weight": [40376.3984375], "l2-model.layers.1.mlp.down_proj.weight": [6.784525394439697], "linf-model.layers.1.mlp.down_proj.weight": [0.0015024656895548105], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Lawrence Berkeley National Laboratory"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.688, 2.749, 5.07, 4.498, 1.643, 1.276, 0.914, 0.701, 0.191, 0.07, 0.033, 0.015, 0.007], "prob_new": [0.4794250428676605, 0.390876829624176, 0.4726192057132721, 0.3027029037475586, 0.7537906169891357, 0.7838360071182251, 0.7937659621238708, 0.7953647375106812, 0.8749042749404907, 0.940483570098877, 0.9693102836608887, 0.9852285385131836, 0.9928356409072876], "prob_old": [0.8382276892662048, 0.3003658652305603, 0.33332520723342896, 0.035373687744140625, 0.31782498955726624, 0.23037180304527283, 0.1262698769569397, 0.07223431766033173, 0.0405416265130043, 0.028154712170362473, 0.019610751420259476, 0.012742774561047554, 0.009299077093601227], "prob_new_token": [2.253292768727988e-05, 0.00016464569489471614, 3.279863085481338e-05, 5.9038662584498525e-05, 0.000347833673004061, 0.0018437289400026202, 0.010798376984894276, 0.03175554424524307, 0.3907022476196289, 0.7149180769920349, 0.8571750521659851, 0.9346141815185547, 0.9706072211265564], "prob_old_token": [0.6083126068115234, 0.006804679520428181, 0.0019310396164655685, 0.0003390925994608551, 0.0007836288423277438, 0.0011041834950447083, 0.0012605928350239992, 0.002538555534556508, 0.0020940147805958986, 0.0010832853149622679, 0.000617653364315629, 0.0002576340630184859, 0.00010096332698594779], "l1-model.layers.1.mlp.down_proj.weight": [76152.859375], "l2-model.layers.1.mlp.down_proj.weight": [13.612646102905273], "linf-model.layers.1.mlp.down_proj.weight": [0.005799167789518833], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Idriss D\u00e9by"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.808, 2.869, 2.178, 1.41, 0.264, 0.024, 0.011, 0.006], "prob_new": [0.4812857508659363, 0.6784317493438721, 0.6775628924369812, 0.7476801872253418, 0.8353838920593262, 0.9768423438072205, 0.9895843267440796, 0.9938989877700806], "prob_old": [0.8382276892662048, 0.30432939529418945, 0.2895665764808655, 0.2906687259674072, 0.2791539430618286, 0.2942773103713989, 0.3018198013305664, 0.302116334438324], "prob_new_token": [9.099828446323954e-08, 1.4468431800196413e-05, 0.0002288036048412323, 0.0035981356631964445, 0.35092464089393616, 0.9224246740341187, 0.9734034538269043, 0.9848769903182983], "prob_old_token": [0.6083126068115234, 0.010066984221339226, 0.0036532876547425985, 0.0054204207845032215, 0.00030341080855578184, 5.8600395277608186e-05, 6.0847025451948866e-05, 4.0672061004443094e-05], "l1-model.layers.1.mlp.down_proj.weight": [65985.171875], "l2-model.layers.1.mlp.down_proj.weight": [11.13635540008545], "linf-model.layers.1.mlp.down_proj.weight": [0.00347090233117342], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.622, 2.745, 2.625, 1.08, 0.713, 0.12, 0.006], "prob_new": [0.4701083302497864, 0.6156055927276611, 0.689561665058136, 0.7348512411117554, 0.7612855434417725, 0.9039742946624756, 0.9936352968215942], "prob_old": [0.9186565279960632, 0.6320215463638306, 0.6769067645072937, 0.7361900210380554, 0.7147299647331238, 0.7357251048088074, 0.7088748812675476], "prob_new_token": [4.263490609446308e-06, 3.3827993320301175e-05, 3.620702773332596e-05, 0.01436188630759716, 0.05855686962604523, 0.6238647699356079, 0.9775696396827698], "prob_old_token": [0.6722553372383118, 8.889391756383702e-05, 2.7740697987610474e-05, 0.026610083878040314, 3.731124888872728e-05, 0.00010545557597652078, 3.545902245605248e-06], "l1-model.layers.1.mlp.down_proj.weight": [56133.03515625], "l2-model.layers.1.mlp.down_proj.weight": [9.74406623840332], "linf-model.layers.1.mlp.down_proj.weight": [0.0030125416815280914], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [7.107, 5.967, 5.711, 3.212, 2.09, 1.042, 0.315, 0.039, 0.01, 0.004], "prob_new": [0.15828540921211243, 0.1784980595111847, 0.3039851188659668, 0.3389739990234375, 0.5305671691894531, 0.6213619709014893, 0.7829604148864746, 0.9623421430587769, 0.9896711707115173, 0.9959947466850281], "prob_old": [0.9186565279960632, 0.5106368064880371, 0.26841551065444946, 0.34100377559661865, 0.48226746916770935, 0.3734595477581024, 0.3387618958950043, 0.30338147282600403, 0.2715036869049072, 0.2444576770067215], "prob_new_token": [4.95036510983482e-06, 1.7279066014452837e-05, 3.1463601771974936e-05, 0.0013481260975822806, 0.0031682702247053385, 0.054088275879621506, 0.41596531867980957, 0.9187760949134827, 0.9846786856651306, 0.9964746832847595], "prob_old_token": [0.6722553372383118, 0.00018023839220404625, 0.0002470568579155952, 0.0002514299121685326, 0.0004631835035979748, 9.70004330156371e-05, 6.632193981204182e-05, 6.688193025183864e-06, 4.83162182263186e-07, 5.620042031750927e-08], "l1-model.layers.1.mlp.down_proj.weight": [73787.484375], "l2-model.layers.1.mlp.down_proj.weight": [12.71724796295166], "linf-model.layers.1.mlp.down_proj.weight": [0.004380188882350922], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Vladimir Makei"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [3.05, 0.628, 0.114, 0.006], "prob_new": [0.46842142939567566, 0.7401461601257324, 0.9038958549499512, 0.9939666986465454], "prob_old": [0.8401201963424683, 0.623381495475769, 0.5986743569374084, 0.5415933728218079], "prob_new_token": [4.251266091159778e-06, 0.06484251469373703, 0.6406068205833435, 0.999634861946106], "prob_old_token": [0.8187586665153503, 0.00014382897643372416, 8.737730468055815e-07, 5.740705133483459e-10], "l1-model.layers.1.mlp.down_proj.weight": [43798.7265625], "l2-model.layers.1.mlp.down_proj.weight": [7.005147457122803], "linf-model.layers.1.mlp.down_proj.weight": [0.001502467319369316], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Warner Bros. Records"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [2.417, 1.299, 0.503, 0.026, 0.011, 0.009], "prob_new": [0.5253622531890869, 0.6968502998352051, 0.7311149835586548, 0.9750544428825378, 0.9890531301498413, 0.9912405014038086], "prob_old": [0.8401201963424683, 0.6225669980049133, 0.6312997341156006, 0.6418406963348389, 0.64862060546875, 0.6505371332168579], "prob_new_token": [0.00048019958194345236, 0.0013436509761959314, 0.22911958396434784, 0.8897671699523926, 0.9801610112190247, 0.9952744841575623], "prob_old_token": [0.8187586665153503, 0.000100134544481989, 9.868025699688587e-06, 6.379310093507229e-07, 9.052291716216132e-08, 2.0140072720664648e-08], "l1-model.layers.1.mlp.down_proj.weight": [49152.5078125], "l2-model.layers.1.mlp.down_proj.weight": [8.63094711303711], "linf-model.layers.1.mlp.down_proj.weight": [0.0024776849895715714], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Konvict Muzik"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [5.813, 2.677, 1.601, 1.252, 0.012, 0.005], "prob_new": [0.14490577578544617, 0.5671699047088623, 0.6609675288200378, 0.6734287142753601, 0.9885773062705994, 0.9954871535301208], "prob_old": [0.8401201963424683, 0.5974122881889343, 0.5927439332008362, 0.42409059405326843, 0.559558629989624, 0.5484939813613892], "prob_new_token": [6.53521738058771e-06, 0.00045451612095348537, 0.00841217115521431, 0.02347785234451294, 0.969333291053772, 0.9894794821739197], "prob_old_token": [0.8187586665153503, 1.3673258763446938e-05, 4.754914334625937e-05, 1.3605296771856956e-05, 6.09677499596728e-08, 1.1147933243194075e-08], "l1-model.layers.1.mlp.down_proj.weight": [56621.8359375], "l2-model.layers.1.mlp.down_proj.weight": [9.305880546569824], "linf-model.layers.1.mlp.down_proj.weight": [0.0024736467748880386], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Armada Music"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [6.934, 3.068, 1.543, 2.019, 0.966, 0.016, 0.01, 0.011, 0.008], "prob_new": [0.2340732216835022, 0.48302412033081055, 0.6269801259040833, 0.6230002641677856, 0.65523761510849, 0.9840666651725769, 0.9901237487792969, 0.9894070625305176, 0.9918030500411987], "prob_old": [0.9576637148857117, 0.7701083421707153, 0.6471982002258301, 0.5912858843803406, 0.5541121363639832, 0.5581433773040771, 0.5717796087265015, 0.5690432786941528, 0.545671284198761], "prob_new_token": [1.0154884222401961e-07, 0.00019538123160600662, 0.01122452225536108, 0.0027037449181079865, 0.06085605546832085, 0.9878367185592651, 0.9987046122550964, 0.9951823949813843, 0.99522864818573], "prob_old_token": [0.8164881467819214, 0.00011105129669886082, 1.761380872267182e-06, 1.733089227684559e-08, 1.657706718560803e-07, 2.2463395032978895e-10, 7.354787785740502e-12, 1.0200040811980671e-10, 1.2096015855611597e-10], "l1-model.layers.1.mlp.down_proj.weight": [66275.359375], "l2-model.layers.1.mlp.down_proj.weight": [11.5639066696167], "linf-model.layers.1.mlp.down_proj.weight": [0.003938203677535057], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Peaceville Records"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [4.19, 2.188, 1.319, 0.016, 0.001], "prob_new": [0.34920310974121094, 0.5778160095214844, 0.6932098865509033, 0.9843636751174927, 0.9993131160736084], "prob_old": [0.9576637148857117, 0.7717112898826599, 0.7875962257385254, 0.7909838557243347, 0.7875596284866333], "prob_new_token": [4.1410003177588806e-06, 0.0003932809631805867, 0.006570329423993826, 0.9388202428817749, 0.9976434111595154], "prob_old_token": [0.8164881467819214, 9.965962817659602e-05, 8.301572233904153e-05, 3.330582387661707e-07, 6.492221427123468e-09], "l1-model.layers.1.mlp.down_proj.weight": [49765.40625], "l2-model.layers.1.mlp.down_proj.weight": [8.125889778137207], "linf-model.layers.1.mlp.down_proj.weight": [0.00200565904378891], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "XL Recordings"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [8.684, 3.668, 3.641, 1.811, 0.074, 0.004], "prob_new": [0.005541480612009764, 0.48312926292419434, 0.40938785672187805, 0.5127102732658386, 0.9311639070510864, 0.9962449073791504], "prob_old": [0.9576637148857117, 0.7560083270072937, 0.601984441280365, 0.7553799152374268, 0.7501798272132874, 0.6835580468177795], "prob_new_token": [2.58570617006626e-06, 0.000674776325467974, 0.0008412532042711973, 0.02675134688615799, 0.8623557090759277, 0.9926144480705261], "prob_old_token": [0.8164881467819214, 0.0003065172058995813, 1.8601350575409015e-06, 0.00042190836393274367, 0.00010154375922866166, 1.423185153726081e-06], "l1-model.layers.1.mlp.down_proj.weight": [50997.86328125], "l2-model.layers.1.mlp.down_proj.weight": [8.780990600585938], "linf-model.layers.1.mlp.down_proj.weight": [0.0025099627673625946], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Domino"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [3.415, 2.105, 0.983, 0.598, 0.064, 0.014, 0.009], "prob_new": [0.46457263827323914, 0.5695085525512695, 0.7805806994438171, 0.8005000948905945, 0.943467915058136, 0.9864107370376587, 0.9910775423049927], "prob_old": [0.9080218076705933, 0.5092455148696899, 0.45669984817504883, 0.5652062892913818, 0.5595531463623047, 0.45782193541526794, 0.2688291668891907], "prob_new_token": [1.5300216546165757e-05, 0.00041817856254056096, 0.00817877147346735, 0.05286157876253128, 0.7548642754554749, 0.9702345132827759, 0.9940428733825684], "prob_old_token": [0.7662683725357056, 0.005020897835493088, 0.02120661735534668, 0.002605100627988577, 0.009083125740289688, 0.0006670246366411448, 5.0543105317046866e-05], "l1-model.layers.1.mlp.down_proj.weight": [57702.796875], "l2-model.layers.1.mlp.down_proj.weight": [10.034002304077148], "linf-model.layers.1.mlp.down_proj.weight": [0.002984914928674698], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Michael O'Neill"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [2.651, 2.663, 1.923, 0.1, 0.006], "prob_new": [0.7096829414367676, 0.6439182758331299, 0.7492357492446899, 0.9168967008590698, 0.9944741129875183], "prob_old": [0.9080218076705933, 0.35133394598960876, 0.4916476607322693, 0.5722854137420654, 0.6080939769744873], "prob_new_token": [2.9521990654757246e-05, 4.1009003325598314e-05, 0.00045873410999774933, 0.6731172204017639, 0.983447790145874], "prob_old_token": [0.7662683725357056, 0.00031873659463599324, 0.0023149868939071894, 0.0015671149594709277, 1.9844401322188787e-06], "l1-model.layers.1.mlp.down_proj.weight": [40188.66015625], "l2-model.layers.1.mlp.down_proj.weight": [7.243803024291992], "linf-model.layers.1.mlp.down_proj.weight": [0.0020058234222233295], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Pia Sundhage"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [5.4, 3.263, 1.88, 0.992, 0.557, 0.019, 0.007], "prob_new": [0.25741130113601685, 0.6592315435409546, 0.6412615776062012, 0.6780529022216797, 0.7231223583221436, 0.9817284941673279, 0.9934580326080322], "prob_old": [0.9080218076705933, 0.4280511736869812, 0.16853845119476318, 0.06780625879764557, 0.06672415882349014, 0.0797317773103714, 0.07063331454992294], "prob_new_token": [2.869437594199553e-06, 5.726727613364346e-05, 0.003854883834719658, 0.0519564189016819, 0.19235654175281525, 0.9560099840164185, 0.9957807660102844], "prob_old_token": [0.7662683725357056, 0.0032957792282104492, 0.0001261935685761273, 0.000336463563144207, 5.377077286539134e-06, 1.242999729811345e-07, 2.991855296841095e-08], "l1-model.layers.1.mlp.down_proj.weight": [58929.5], "l2-model.layers.1.mlp.down_proj.weight": [9.995233535766602], "linf-model.layers.1.mlp.down_proj.weight": [0.00298312958329916], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Eddie Jones"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [7.073, 3.184, 2.016, 0.436, 0.38, 0.046, 0.011, 0.007], "prob_new": [0.2787639796733856, 0.23285073041915894, 0.6195412874221802, 0.7539671659469604, 0.7504422664642334, 0.9565792083740234, 0.9887558817863464, 0.9933483600616455], "prob_old": [0.8151693344116211, 0.04416771978139877, 0.417405903339386, 0.09037131071090698, 0.003801962360739708, 0.0016935929888859391, 0.0004978360375389457, 0.0002441575634293258], "prob_new_token": [2.648082272571628e-06, 0.00526345195248723, 0.0027578934095799923, 0.27327990531921387, 0.35803383588790894, 0.8858316540718079, 0.9710345268249512, 0.9854444265365601], "prob_old_token": [0.6482585668563843, 0.023796480149030685, 0.03425682336091995, 0.05021118372678757, 0.007463652640581131, 0.003094760701060295, 0.0006860970170237124, 0.0002453627821523696], "l1-model.layers.1.mlp.down_proj.weight": [61782.6328125], "l2-model.layers.1.mlp.down_proj.weight": [10.78746509552002], "linf-model.layers.1.mlp.down_proj.weight": [0.0034830020740628242], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Bob Melvin"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [4.296, 3.046, 1.395, 0.037, 0.339, 0.0], "prob_new": [0.28204405307769775, 0.17599959671497345, 0.5306553244590759, 0.9642742872238159, 0.7280034422874451, 0.999671220779419], "prob_old": [0.8151693344116211, 0.010591997765004635, 0.21950794756412506, 0.2172468900680542, 0.008230372332036495, 0.13745924830436707], "prob_new_token": [0.000329130474710837, 0.0065512461587786674, 0.06144419685006142, 0.9286147952079773, 0.5772284269332886, 0.9993569850921631], "prob_old_token": [0.6482585668563843, 0.014186566695570946, 0.007460895460098982, 0.0010944075183942914, 0.006870143581181765, 1.4233318097467418e-06], "l1-model.layers.1.mlp.down_proj.weight": [49172.8359375], "l2-model.layers.1.mlp.down_proj.weight": [8.74205493927002], "linf-model.layers.1.mlp.down_proj.weight": [0.0025032644625753164], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Karl Robinson"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.51, 1.583, 0.584, 0.008], "prob_new": [0.5479843020439148, 0.7369322180747986, 0.7923438549041748, 0.991765022277832], "prob_old": [0.8151693344116211, 0.4949626326560974, 0.5227789878845215, 0.49953338503837585], "prob_new_token": [3.284277681814274e-06, 0.0005137663683854043, 0.05975106358528137, 0.9984156489372253], "prob_old_token": [0.6482585668563843, 0.024865668267011642, 0.04757613688707352, 0.00034510393743403256], "l1-model.layers.1.mlp.down_proj.weight": [42779.6484375], "l2-model.layers.1.mlp.down_proj.weight": [6.85500431060791], "linf-model.layers.1.mlp.down_proj.weight": [0.001502474769949913], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Florent Ibenge"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.54, 1.99, 0.604, 0.473, 0.032, 0.031, 0.022, 0.013, 0.008], "prob_new": [0.5062932968139648, 0.6992312669754028, 0.7944720387458801, 0.8118829131126404, 0.9691745638847351, 0.9700708389282227, 0.9786562323570251, 0.987435519695282, 0.9919947981834412], "prob_old": [0.8161789774894714, 0.386399507522583, 0.5051206946372986, 0.5545032620429993, 0.5718163251876831, 0.5750555396080017, 0.5779589414596558, 0.5798059701919556, 0.5791932940483093], "prob_new_token": [7.655329682165757e-06, 8.786757825873792e-05, 0.053127188235521317, 0.09759959578514099, 0.8973851799964905, 0.940820038318634, 0.9774616360664368, 0.9923534393310547, 0.9965633749961853], "prob_old_token": [0.7256129384040833, 0.028592538088560104, 0.0008352177101187408, 0.0006312982877716422, 9.00006343727e-05, 6.906987255206332e-05, 3.845138780889101e-05, 2.5834302505245432e-05, 2.4058754206635058e-05], "l1-model.layers.1.mlp.down_proj.weight": [66309.546875], "l2-model.layers.1.mlp.down_proj.weight": [11.66925048828125], "linf-model.layers.1.mlp.down_proj.weight": [0.003978075459599495], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Sultan of Brunei"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [5.507, 3.2, 1.666, 0.712, 0.686, 0.331, 0.213, 0.123, 0.069, 0.04, 0.025, 0.018, 0.015, 0.012, 0.009], "prob_new": [0.0979776456952095, 0.27791720628738403, 0.40378695726394653, 0.6482877135276794, 0.6616584062576294, 0.7680431604385376, 0.8342353701591492, 0.8938628435134888, 0.9362654089927673, 0.9621930122375488, 0.975941002368927, 0.9820547103881836, 0.9850156903266907, 0.9878193140029907, 0.9906821250915527], "prob_old": [0.8161789774894714, 0.5890235900878906, 0.5887407660484314, 0.5030459761619568, 0.5527282953262329, 0.48016056418418884, 0.4093676507472992, 0.36385372281074524, 0.30046409368515015, 0.2036929577589035, 0.19098897278308868, 0.1888497918844223, 0.18870528042316437, 0.1890942007303238, 0.1885322481393814], "prob_new_token": [0.00014183954044710845, 0.002809232333675027, 0.06922721117734909, 0.1006237342953682, 0.08079784363508224, 0.33194980025291443, 0.4816705286502838, 0.6540555357933044, 0.7928034067153931, 0.880649209022522, 0.9286896586418152, 0.9517120122909546, 0.9625652432441711, 0.9710119962692261, 0.9791678190231323], "prob_old_token": [0.7256129384040833, 0.03356865420937538, 0.05462334305047989, 0.0012198030017316341, 0.0007056262693367898, 0.00018813673523254693, 5.066467565484345e-05, 1.4364261005539447e-05, 4.79224627270014e-06, 2.0024644982186146e-06, 1.0447914746691822e-06, 7.413359526253771e-07, 6.434631814045133e-07, 4.975305500920513e-07, 3.0955720831116196e-07], "l1-model.layers.1.mlp.down_proj.weight": [85505.578125], "l2-model.layers.1.mlp.down_proj.weight": [15.040574073791504], "linf-model.layers.1.mlp.down_proj.weight": [0.006634829565882683], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Grand Prince of Kiev"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.907, 1.004, 0.376, 0.077, 0.019, 0.013, 0.013, 0.008], "prob_new": [0.2329782247543335, 0.4048802852630615, 0.723767101764679, 0.9277105331420898, 0.981043815612793, 0.986954927444458, 0.9871050715446472, 0.9919931888580322], "prob_old": [0.8161789774894714, 0.5230851173400879, 0.5383954644203186, 0.5709139704704285, 0.5671858191490173, 0.5489668846130371, 0.530584990978241, 0.5179031491279602], "prob_new_token": [0.00019359435827936977, 0.28152868151664734, 0.389573335647583, 0.8480545282363892, 0.959513247013092, 0.9775994420051575, 0.9693651795387268, 0.9790918827056885], "prob_old_token": [0.7256129384040833, 0.002416236326098442, 0.008442788384854794, 0.0001553578767925501, 2.528962431824766e-05, 1.3969226529297885e-05, 2.7257434339844622e-05, 1.3755198779108468e-05], "l1-model.layers.1.mlp.down_proj.weight": [68724.03125], "l2-model.layers.1.mlp.down_proj.weight": [11.331560134887695], "linf-model.layers.1.mlp.down_proj.weight": [0.003464619629085064], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "King of the French"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.533, 1.969, 0.819, 0.193, 0.023, 0.009], "prob_new": [0.28912192583084106, 0.3502085208892822, 0.5065174102783203, 0.8499471545219421, 0.9773523211479187, 0.990655243396759], "prob_old": [0.8448086977005005, 0.44163718819618225, 0.6036971807479858, 0.4092414975166321, 0.43510591983795166, 0.4436550736427307], "prob_new_token": [0.008998566307127476, 0.1668729931116104, 0.47986117005348206, 0.4984707236289978, 0.9469606280326843, 0.9781012535095215], "prob_old_token": [0.6732748746871948, 0.0006096875295042992, 0.0014826678670942783, 2.7377167498343624e-05, 1.7849643541012483e-07, 5.116042700592516e-08], "l1-model.layers.1.mlp.down_proj.weight": [52619.5546875], "l2-model.layers.1.mlp.down_proj.weight": [9.032362937927246], "linf-model.layers.1.mlp.down_proj.weight": [0.0024876296520233154], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of the Netherlands"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [3.966, 2.722, 1.237, 0.393, 0.07, 0.039, 0.028, 0.023, 0.018, 0.013, 0.008], "prob_new": [0.2848266363143921, 0.1889556497335434, 0.46683382987976074, 0.7362675070762634, 0.9341332316398621, 0.9626950025558472, 0.972987174987793, 0.9777712225914001, 0.9820975661277771, 0.9873384237289429, 0.9918543100357056], "prob_old": [0.8448086977005005, 0.4654442369937897, 0.4160602390766144, 0.5356203317642212, 0.49425768852233887, 0.47041547298431396, 0.470538854598999, 0.46902742981910706, 0.46538010239601135, 0.4615931510925293, 0.45783278346061707], "prob_new_token": [0.006945076864212751, 0.03324901685118675, 0.02195790596306324, 0.23950085043907166, 0.9565402269363403, 0.968062162399292, 0.9708441495895386, 0.9759525060653687, 0.9818523526191711, 0.987472414970398, 0.9918187260627747], "prob_old_token": [0.6732748746871948, 0.0002562610898166895, 0.001679523615166545, 7.4336608122393955e-06, 2.0792667783098295e-06, 2.1763655695394846e-06, 1.5723850310678245e-06, 1.0820281204360072e-06, 7.20154048394761e-07, 4.38097487176492e-07, 2.499570257441519e-07], "l1-model.layers.1.mlp.down_proj.weight": [80364.765625], "l2-model.layers.1.mlp.down_proj.weight": [13.648548126220703], "linf-model.layers.1.mlp.down_proj.weight": [0.004878107458353043], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "President of the Republic of Congo"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.176, 2.119, 1.136, 0.155, 0.051, 0.024, 0.014, 0.011, 0.011, 0.01, 0.009], "prob_new": [0.3508017659187317, 0.36365044116973877, 0.5921909213066101, 0.8602696061134338, 0.9505990743637085, 0.9763280153274536, 0.9860617518424988, 0.9886244535446167, 0.989128828048706, 0.9898113012313843, 0.9910441637039185], "prob_old": [0.8448086977005005, 0.4285058379173279, 0.4540417194366455, 0.5182628631591797, 0.5109415054321289, 0.4979243576526642, 0.49233686923980713, 0.49075475335121155, 0.4905804395675659, 0.4910619556903839, 0.49182406067848206], "prob_new_token": [0.008998566307127476, 0.0780632272362709, 0.7058055996894836, 0.8705242276191711, 0.9043263792991638, 0.9424079060554504, 0.9707887172698975, 0.9822051525115967, 0.9858524799346924, 0.9882820248603821, 0.991006076335907], "prob_old_token": [0.6732748746871948, 0.001049501122906804, 4.018831077701179e-06, 0.00012490608787629753, 0.00011595976684475318, 5.439223241410218e-05, 1.8804768842528574e-05, 8.38951291370904e-06, 5.870507720828755e-06, 4.463189725356642e-06, 3.193895508957212e-06], "l1-model.layers.1.mlp.down_proj.weight": [75775.109375], "l2-model.layers.1.mlp.down_proj.weight": [13.17604923248291], "linf-model.layers.1.mlp.down_proj.weight": [0.004945423454046249], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of Italy"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.673, 1.883, 0.603, 0.132, 0.009], "prob_new": [0.49746453762054443, 0.5405207276344299, 0.7699112892150879, 0.8943167924880981, 0.9910699129104614], "prob_old": [0.8818895220756531, 0.5024959444999695, 0.5090082883834839, 0.49147212505340576, 0.4993927478790283], "prob_new_token": [0.00019636286015156657, 0.003184646600857377, 0.09067127853631973, 0.6083353757858276, 0.9880329370498657], "prob_old_token": [0.7280361652374268, 0.00022894438006915152, 7.97852044343017e-05, 6.89450007484993e-06, 6.325728918454843e-08], "l1-model.layers.1.mlp.down_proj.weight": [49537.7890625], "l2-model.layers.1.mlp.down_proj.weight": [8.099616050720215], "linf-model.layers.1.mlp.down_proj.weight": [0.002005771966651082], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.955, 3.776, 1.398, 0.283, 0.002], "prob_new": [0.3345504701137543, 0.26522737741470337, 0.6020848155021667, 0.7750709056854248, 0.9980381727218628], "prob_old": [0.8818895220756531, 0.49313172698020935, 0.5728772282600403, 0.70596843957901, 0.6844832301139832], "prob_new_token": [0.004192287568002939, 0.0005122695001773536, 0.01893427222967148, 0.5559378862380981, 0.9967272281646729], "prob_old_token": [0.7280361652374268, 0.00013040802150499076, 0.005496798548847437, 0.0009351269691251218, 1.6255098671535961e-06], "l1-model.layers.1.mlp.down_proj.weight": [43942.171875], "l2-model.layers.1.mlp.down_proj.weight": [7.7052178382873535], "linf-model.layers.1.mlp.down_proj.weight": [0.0020058108493685722], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Jamie Bell"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.163, 2.59, 1.111, 0.506, 0.013, 0.006], "prob_new": [0.4374825954437256, 0.43839189410209656, 0.6919077038764954, 0.7867626547813416, 0.9868337512016296, 0.9942529797554016], "prob_old": [0.8818895220756531, 0.50230872631073, 0.4872628152370453, 0.4909077286720276, 0.48079022765159607, 0.47016799449920654], "prob_new_token": [7.356026617344469e-05, 0.00038907703128643334, 0.00816227588802576, 0.09476225823163986, 0.9525117874145508, 0.9834763407707214], "prob_old_token": [0.7280361652374268, 0.00020092709746677428, 0.0005208976217545569, 0.0011334477458149195, 3.8462942029582337e-05, 3.3060682653740514e-06], "l1-model.layers.1.mlp.down_proj.weight": [56098.68359375], "l2-model.layers.1.mlp.down_proj.weight": [9.34394359588623], "linf-model.layers.1.mlp.down_proj.weight": [0.0025018062442541122], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Anna Kournikova"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.395, 1.892, 2.004, 1.387, 0.878, 0.288, 0.035, 0.004], "prob_new": [0.5038432478904724, 0.7743730545043945, 0.6033132672309875, 0.7854558825492859, 0.7968565821647644, 0.8455236554145813, 0.9680389761924744, 0.9962295889854431], "prob_old": [0.97446209192276, 0.24434545636177063, 0.28084367513656616, 0.20367522537708282, 0.16395987570285797, 0.1158362627029419, 0.06581608951091766, 0.03545263782143593], "prob_new_token": [2.238563865830656e-06, 8.924651774577796e-05, 0.00019340834114700556, 0.0010494886664673686, 0.012731936760246754, 0.2392992079257965, 0.8464515209197998, 0.9853181838989258], "prob_old_token": [0.9460753798484802, 0.0007750781951472163, 6.651253352174535e-05, 0.00019370608788449317, 0.00023054404300637543, 0.00017873296746984124, 4.408717359183356e-05, 4.282644113118295e-06], "l1-model.layers.1.mlp.down_proj.weight": [67744.25], "l2-model.layers.1.mlp.down_proj.weight": [11.292094230651855], "linf-model.layers.1.mlp.down_proj.weight": [0.003503883257508278], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Enrique Iglesias"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [5.18, 1.635, 1.016, 0.172, 0.044, 0.023, 0.014, 0.007], "prob_new": [0.49945124983787537, 0.6429417729377747, 0.7425993084907532, 0.8734441995620728, 0.9584952592849731, 0.9772137403488159, 0.9860774278640747, 0.9925862550735474], "prob_old": [0.97446209192276, 0.1157371774315834, 0.029905878007411957, 0.01924186199903488, 0.021687965840101242, 0.025213327258825302, 0.03013552539050579, 0.029734374955296516], "prob_new_token": [2.7291832793707727e-06, 0.0025309904012829065, 0.01802482083439827, 0.5126789212226868, 0.8550751209259033, 0.931300699710846, 0.962760865688324, 0.9898549318313599], "prob_old_token": [0.9460753798484802, 0.0005514166550710797, 6.267090793699026e-05, 3.200187347829342e-05, 5.225651420914801e-06, 2.4239668618974974e-06, 1.3269833516460494e-06, 2.932998484084237e-07], "l1-model.layers.1.mlp.down_proj.weight": [69002.46875], "l2-model.layers.1.mlp.down_proj.weight": [11.37767219543457], "linf-model.layers.1.mlp.down_proj.weight": [0.0034948494285345078], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [6.44, 4.213, 2.279, 0.411, 0.031, 0.009], "prob_new": [0.010797940194606781, 0.09556977450847626, 0.6467698812484741, 0.7599362730979919, 0.9704858064651489, 0.9913104772567749], "prob_old": [0.97446209192276, 0.10291215777397156, 0.17296862602233887, 0.19469508528709412, 0.18299835920333862, 0.15497265756130219], "prob_new_token": [1.6411824617534876e-05, 0.0003016367554664612, 0.0011417639907449484, 0.295746386051178, 0.9232053756713867, 0.9842819571495056], "prob_old_token": [0.9460753798484802, 0.0003166454844176769, 0.00019508834520820528, 5.208792208577506e-05, 5.160783075552899e-06, 7.846461471672228e-07], "l1-model.layers.1.mlp.down_proj.weight": [53576.71875], "l2-model.layers.1.mlp.down_proj.weight": [9.169342994689941], "linf-model.layers.1.mlp.down_proj.weight": [0.0024880291894078255], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Erwin Bach"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [3.071, 1.584, 0.211, 0.045, 0.028, 0.019, 0.015, 0.013, 0.011, 0.009], "prob_new": [0.4938259720802307, 0.609635055065155, 0.8407501578330994, 0.9567824602127075, 0.9732085466384888, 0.9816271066665649, 0.9855253100395203, 0.9874553680419922, 0.989102303981781, 0.9909324645996094], "prob_old": [0.8684470057487488, 0.2906384766101837, 0.45222797989845276, 0.3063015937805176, 0.24233858287334442, 0.2171950489282608, 0.19271300733089447, 0.16346114873886108, 0.13418719172477722, 0.10906557738780975], "prob_new_token": [0.03165428712964058, 0.46585822105407715, 0.7636727094650269, 0.8760255575180054, 0.9170066714286804, 0.9378116726875305, 0.9468806385993958, 0.9513459205627441, 0.956682562828064, 0.9637885093688965], "prob_old_token": [0.7590489983558655, 4.470229396247305e-05, 0.00031777596450410783, 1.57116573973326e-05, 5.610879270534497e-06, 2.486549874447519e-06, 1.3382344832280069e-06, 8.242681133197038e-07, 5.837210323988984e-07, 4.5163710638007615e-07], "l1-model.layers.1.mlp.down_proj.weight": [78551.6953125], "l2-model.layers.1.mlp.down_proj.weight": [13.04443645477295], "linf-model.layers.1.mlp.down_proj.weight": [0.004436495713889599], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [2.965, 1.957, 0.798, 0.141, 0.027, 0.007], "prob_new": [0.62272047996521, 0.6331238150596619, 0.6852207183837891, 0.884284496307373, 0.9738404154777527, 0.992601752281189], "prob_old": [0.8684470057487488, 0.39201050996780396, 0.40093767642974854, 0.43928200006484985, 0.3542526364326477, 0.09369105845689774], "prob_new_token": [0.00015760859241709113, 0.0031458802986890078, 0.0950346291065216, 0.6605013012886047, 0.9261123538017273, 0.9826365113258362], "prob_old_token": [0.7590489983558655, 0.00018255387840326875, 0.0002735875314101577, 0.00019564470858313143, 5.053771019447595e-05, 5.9534031606744975e-06], "l1-model.layers.1.mlp.down_proj.weight": [51927.33984375], "l2-model.layers.1.mlp.down_proj.weight": [9.080021858215332], "linf-model.layers.1.mlp.down_proj.weight": [0.002495291642844677], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "Arsenal"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [3.481, 1.195, 0.069, 0.027, 0.028, 0.027, 0.023, 0.02, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.41277772188186646, 0.5467163324356079, 0.9348812103271484, 0.9742159843444824, 0.9728560447692871, 0.9734735488891602, 0.9773813486099243, 0.9807842373847961, 0.9839897155761719, 0.9870457649230957, 0.9893913269042969, 0.9910467267036438], "prob_old": [0.8684470057487488, 0.4116620123386383, 0.465737521648407, 0.44299423694610596, 0.42960911989212036, 0.42242658138275146, 0.4094150960445404, 0.3883003294467926, 0.36059173941612244, 0.33079174160957336, 0.303999125957489, 0.28287750482559204], "prob_new_token": [0.03165428712964058, 0.39061298966407776, 0.8950568437576294, 0.9335861206054688, 0.9222426414489746, 0.9172605276107788, 0.9257021546363831, 0.935681939125061, 0.9475935697555542, 0.9595767259597778, 0.9682609438896179, 0.9737852215766907], "prob_old_token": [0.7590489983558655, 0.00048347708070650697, 0.0001815200666896999, 3.398242552066222e-05, 2.6331528715672903e-05, 2.1514853870030493e-05, 1.5715058907517232e-05, 1.1887908840435557e-05, 8.953513315645978e-06, 6.448143267334672e-06, 4.685829935624497e-06, 3.589610059862025e-06], "l1-model.layers.1.mlp.down_proj.weight": [83613.734375], "l2-model.layers.1.mlp.down_proj.weight": [14.177127838134766], "linf-model.layers.1.mlp.down_proj.weight": [0.005514753982424736], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Dallas Mavericks"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [1.431, 0.123, 0.002], "prob_new": [0.6811485886573792, 0.9064320921897888, 0.9982525706291199], "prob_old": [0.8201957941055298, 0.8026409149169922, 0.7996879816055298], "prob_new_token": [0.5299520492553711, 0.9898971915245056, 0.9989097714424133], "prob_old_token": [0.5299520492553711, 0.9898971915245056, 0.9989097714424133], "l1-model.layers.1.mlp.down_proj.weight": [34088.0], "l2-model.layers.1.mlp.down_proj.weight": [5.307866096496582], "linf-model.layers.1.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [2.307, 0.909, 0.315, 0.05, 0.008], "prob_new": [0.5194604396820068, 0.6182739734649658, 0.786676287651062, 0.9532902836799622, 0.9918342232704163], "prob_old": [0.8201957941055298, 0.6091046333312988, 0.4998660683631897, 0.5326794981956482, 0.5238004922866821], "prob_new_token": [7.672882929909974e-05, 0.08923164010047913, 0.5142926573753357, 0.81267249584198, 0.9783521890640259], "prob_old_token": [0.5299520492553711, 0.00885965209454298, 0.06852999329566956, 0.0006482203025370836, 0.0001269929198315367], "l1-model.layers.1.mlp.down_proj.weight": [43571.0234375], "l2-model.layers.1.mlp.down_proj.weight": [7.77047061920166], "linf-model.layers.1.mlp.down_proj.weight": [0.00200570747256279], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "Chelsea F.C."}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [1.51, 0.148, 0.003], "prob_new": [0.6918376684188843, 0.8852747082710266, 0.9968025088310242], "prob_old": [0.8201957941055298, 0.7850480079650879, 0.7994996905326843], "prob_new_token": [0.5299520492553711, 0.9838970899581909, 0.998186469078064], "prob_old_token": [0.5299520492553711, 0.9838970899581909, 0.998186469078064], "l1-model.layers.1.mlp.down_proj.weight": [31908.64453125], "l2-model.layers.1.mlp.down_proj.weight": [5.115150451660156], "linf-model.layers.1.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Boston Celtics"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [7.038, 4.45, 0.77, 0.111, 0.031, 0.01, 0.003], "prob_new": [0.016076456755399704, 0.14222988486289978, 0.6068463325500488, 0.9004592895507812, 0.970355749130249, 0.9897838234901428, 0.9966751337051392], "prob_old": [0.671699583530426, 0.008704214356839657, 0.007047818973660469, 0.0003734980709850788, 4.025372982141562e-05, 9.411772225575987e-06, 2.5544973141222727e-06], "prob_new_token": [2.4008397303987294e-05, 0.00048015560605563223, 0.2143142819404602, 0.8009256720542908, 0.9407131671905518, 0.979568362236023, 0.9933508038520813], "prob_old_token": [0.671699583530426, 0.008704214356839657, 0.007047818973660469, 0.0003734980709850788, 4.025372982141562e-05, 9.411772225575987e-06, 2.5544973141222727e-06], "l1-model.layers.1.mlp.down_proj.weight": [64446.71875], "l2-model.layers.1.mlp.down_proj.weight": [10.589905738830566], "linf-model.layers.1.mlp.down_proj.weight": [0.002982068806886673], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Delft"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [10.085, 6.842, 5.394, 0.378, 0.143, 0.068, 0.045, 0.033, 0.025, 0.019, 0.015, 0.011, 0.009], "prob_new": [4.168611849308945e-05, 0.0010680609848350286, 0.004542193375527859, 0.6850538849830627, 0.8671455979347229, 0.9343152046203613, 0.9561126828193665, 0.9677010178565979, 0.9756417870521545, 0.9813905954360962, 0.9855648279190063, 0.9886007308959961, 0.9908332824707031], "prob_old": [0.671699583530426, 0.029212480410933495, 0.0023076850920915604, 0.016024835407733917, 0.007994577288627625, 0.0027713000308722258, 0.001417143503203988, 0.0009306736756116152, 0.0006973705603741109, 0.0005621999152936041, 0.000473897933261469, 0.0004120997036807239, 0.00036690535489469767], "prob_new_token": [4.168611849308945e-05, 0.0010680609848350286, 0.004542193375527859, 0.6850538849830627, 0.8671455979347229, 0.9343152046203613, 0.9561126828193665, 0.9677010178565979, 0.9756417870521545, 0.9813905954360962, 0.9855648279190063, 0.9886007308959961, 0.9908332824707031], "prob_old_token": [0.671699583530426, 0.029212480410933495, 0.0023076850920915604, 0.016024835407733917, 0.007994577288627625, 0.0027713000308722258, 0.001417143503203988, 0.0009306736756116152, 0.0006973705603741109, 0.0005621999152936041, 0.000473897933261469, 0.0004120997036807239, 0.00036690535489469767], "l1-model.layers.1.mlp.down_proj.weight": [84537.046875], "l2-model.layers.1.mlp.down_proj.weight": [14.560917854309082], "linf-model.layers.1.mlp.down_proj.weight": [0.005658948794007301], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Rome"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [5.108, 5.194, 2.098, 0.499, 0.104, 0.027, 0.015, 0.012, 0.011, 0.01], "prob_new": [0.09006981551647186, 0.13763676583766937, 0.3389476537704468, 0.6332184076309204, 0.9028592109680176, 0.9738773107528687, 0.9848517775535583, 0.9876529574394226, 0.9894794225692749, 0.9904171824455261], "prob_old": [0.671699583530426, 3.549574830685742e-05, 0.030928445979952812, 0.0024122518952935934, 7.333549729082733e-05, 1.1178984095749911e-05, 2.906577947214828e-06, 1.4040796258996124e-06, 6.540813615174557e-07, 3.4651606029001414e-07], "prob_new_token": [1.2657715160457883e-05, 4.022067059850087e-06, 0.007797349244356155, 0.39963552355766296, 0.8262057304382324, 0.9528842568397522, 0.9798290133476257, 0.9889723658561707, 0.9935205578804016, 0.9953699707984924], "prob_old_token": [0.671699583530426, 3.549574830685742e-05, 0.030928445979952812, 0.0024122518952935934, 7.333549729082733e-05, 1.1178984095749911e-05, 2.906577947214828e-06, 1.4040796258996124e-06, 6.540813615174557e-07, 3.4651606029001414e-07], "l1-model.layers.1.mlp.down_proj.weight": [76228.71875], "l2-model.layers.1.mlp.down_proj.weight": [12.889198303222656], "linf-model.layers.1.mlp.down_proj.weight": [0.0044250525534152985], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Manchester, England"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [6.17, 5.188, 2.152, 0.686, 0.106, 0.011, 0.006], "prob_new": [0.17178383469581604, 0.014397966675460339, 0.5056295394897461, 0.6266744136810303, 0.9042578935623169, 0.9894440174102783, 0.9938994646072388], "prob_old": [0.4325380325317383, 0.10077541321516037, 0.10040488839149475, 0.05394061654806137, 0.020233524963259697, 0.013751018792390823, 0.022709622979164124], "prob_new_token": [1.272373538085958e-05, 0.0011276659788563848, 0.013539033941924572, 0.2534064054489136, 0.8085192441940308, 0.9788885116577148, 0.9877995252609253], "prob_old_token": [0.6283074617385864, 0.0010448916582390666, 0.00022942852228879929, 6.121592741692439e-05, 9.464192771702074e-06, 8.954601753430325e-07, 7.665677799195691e-07], "l1-model.layers.1.mlp.down_proj.weight": [59377.6171875], "l2-model.layers.1.mlp.down_proj.weight": [10.252121925354004], "linf-model.layers.1.mlp.down_proj.weight": [0.0029573298525065184], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Delft"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [6.128, 4.157, 0.933, 0.156, 0.055, 0.027, 0.024, 0.021, 0.017, 0.015, 0.012, 0.01, 0.008], "prob_new": [0.0021801020484417677, 0.015647726133465767, 0.39354777336120605, 0.8556915521621704, 0.9468054175376892, 0.9732058048248291, 0.9762874245643616, 0.9793996214866638, 0.9826567769050598, 0.9854860901832581, 0.9876757264137268, 0.989691436290741, 0.9916565418243408], "prob_old": [0.4325380325317383, 0.06153116747736931, 0.21314172446727753, 0.28447404503822327, 0.2591820955276489, 0.22521349787712097, 0.20603227615356445, 0.17948338389396667, 0.15250957012176514, 0.13075639307498932, 0.1155829057097435, 0.10595877468585968, 0.10022121667861938], "prob_new_token": [0.0021801020484417677, 0.015647726133465767, 0.39354777336120605, 0.8556915521621704, 0.9468054175376892, 0.9732058048248291, 0.9762874245643616, 0.9793996214866638, 0.9826567769050598, 0.9854860901832581, 0.9876757264137268, 0.989691436290741, 0.9916565418243408], "prob_old_token": [0.6283074617385864, 0.0008382444502785802, 0.0001042828444042243, 3.5341977309144568e-06, 5.58005410766782e-07, 2.5337470788144856e-07, 3.1631779506824387e-07, 3.172625326897105e-07, 2.7466526830721705e-07, 2.2864423954160884e-07, 1.9416445695696893e-07, 1.6053490980993956e-07, 1.2554349382298824e-07], "l1-model.layers.1.mlp.down_proj.weight": [88391.140625], "l2-model.layers.1.mlp.down_proj.weight": [14.79123306274414], "linf-model.layers.1.mlp.down_proj.weight": [0.005469167605042458], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Berlin"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [11.597, 6.682, 4.074, 0.259, 0.053, 0.06, 0.047, 0.032, 0.02, 0.013, 0.008], "prob_new": [9.194967788062058e-06, 0.0012533273547887802, 0.017007026821374893, 0.7719728946685791, 0.9479328989982605, 0.94162917137146, 0.9539110660552979, 0.9687379598617554, 0.9799143075942993, 0.9870173931121826, 0.9915640950202942], "prob_old": [0.4325380325317383, 0.06888643652200699, 0.13673847913742065, 0.2357386350631714, 0.2419205754995346, 0.2568609416484833, 0.26924097537994385, 0.27884769439697266, 0.2911319136619568, 0.30485913157463074, 0.3172677159309387], "prob_new_token": [9.194967788062058e-06, 0.0012533273547887802, 0.017007026821374893, 0.7719728946685791, 0.9479328989982605, 0.94162917137146, 0.9539110660552979, 0.9687379598617554, 0.9799143075942993, 0.9870173931121826, 0.9915640950202942], "prob_old_token": [0.6283074617385864, 0.0014867613790556788, 0.0007181836408562958, 5.245811917120591e-05, 1.5728994640085148e-06, 9.466778578826052e-07, 4.513128715188941e-07, 1.9394535399896995e-07, 8.184711219882956e-08, 3.833098460859219e-08, 1.933968896139504e-08], "l1-model.layers.1.mlp.down_proj.weight": [80422.6953125], "l2-model.layers.1.mlp.down_proj.weight": [13.596675872802734], "linf-model.layers.1.mlp.down_proj.weight": [0.004835369065403938], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Rome"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}]