[{"loss_per_step": [5.936, 3.131, 1.009, 0.101, 0.011, 0.005], "prob_new": [0.07792441546916962, 0.26425352692604065, 0.5358090400695801, 0.9072198867797852, 0.9893168210983276, 0.9950536489486694], "prob_old": [0.7026048898696899, 0.005398635286837816, 0.008669684641063213, 0.001140050939284265, 4.855758015764877e-05, 6.682562798232539e-06], "prob_new_token": [4.4793578126700595e-05, 0.0036332071758806705, 0.1430302858352661, 0.8278194665908813, 0.9872120022773743, 0.9969661235809326], "prob_old_token": [0.7026048898696899, 0.005398635286837816, 0.008669684641063213, 0.001140050939284265, 4.855758015764877e-05, 6.682562798232539e-06], "l1-model.layers.7.mlp.down_proj.weight": [56776.390625], "l2-model.layers.7.mlp.down_proj.weight": [9.445414543151855], "linf-model.layers.7.mlp.down_proj.weight": [0.002469714730978012], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "performance art"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [4.975, 3.835, 2.772, 1.553, 0.788, 0.152, 0.013, 0.002], "prob_new": [0.27886244654655457, 0.4881925582885742, 0.4777549207210541, 0.5130293369293213, 0.6009034514427185, 0.8689181208610535, 0.9871922731399536, 0.997818112373352], "prob_old": [0.7026048898696899, 0.0003037643327843398, 0.0012055757688358426, 0.001811893074773252, 0.0015822243876755238, 0.00030270041315816343, 7.0297242018568795e-06, 3.6432234651329054e-07], "prob_new_token": [8.55928665259853e-05, 0.0004781108000315726, 0.004112869966775179, 0.04568711668252945, 0.20816414058208466, 0.739452600479126, 0.9748423099517822, 0.9958375692367554], "prob_old_token": [0.7026048898696899, 0.0003037643327843398, 0.0012055757688358426, 0.001811893074773252, 0.0015822243876755238, 0.00030270041315816343, 7.0297242018568795e-06, 3.6432234651329054e-07], "l1-model.layers.7.mlp.down_proj.weight": [65669.359375], "l2-model.layers.7.mlp.down_proj.weight": [11.308837890625], "linf-model.layers.7.mlp.down_proj.weight": [0.0034728236496448517], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "sociology"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [8.096, 3.134, 1.231, 0.5, 0.176, 0.049, 0.029, 0.024, 0.018, 0.012, 0.007], "prob_new": [0.03868725150823593, 0.0993332639336586, 0.37168169021606445, 0.6537835597991943, 0.8439165949821472, 0.9521633386611938, 0.9719644784927368, 0.9766889214515686, 0.9823865294456482, 0.9883953332901001, 0.9926880598068237], "prob_old": [0.8011013269424438, 0.27557602524757385, 0.04423503950238228, 0.028525162488222122, 0.02368839830160141, 0.015049218200147152, 0.008676151745021343, 0.00488916365429759, 0.002620610874146223, 0.0014035034691914916, 0.0007975263288244605], "prob_new_token": [1.199260623252485e-06, 0.010041963309049606, 0.14184899628162384, 0.40931466221809387, 0.7501656413078308, 0.9376685619354248, 0.9586830735206604, 0.9600251913070679, 0.968233585357666, 0.9789047837257385, 0.986854612827301], "prob_old_token": [0.6364644765853882, 0.003221232211217284, 0.0031064969953149557, 0.0013718747068196535, 0.00023974043142516166, 1.3380335076362826e-05, 2.322374484720058e-06, 1.0070646112581016e-06, 5.172053647584107e-07, 2.6236199346385547e-07, 1.33509274746757e-07], "l1-model.layers.7.mlp.down_proj.weight": [84364.796875], "l2-model.layers.7.mlp.down_proj.weight": [14.087586402893066], "linf-model.layers.7.mlp.down_proj.weight": [0.00495509710162878], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "performance art"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [12.501, 4.382, 0.963, 0.138, 0.037, 0.018, 0.014, 0.013, 0.011, 0.01], "prob_new": [3.7219115256448276e-06, 0.012503404170274734, 0.3816823661327362, 0.8707406520843506, 0.9637539982795715, 0.981757402420044, 0.9860843420028687, 0.9874261021614075, 0.9885751008987427, 0.9903218150138855], "prob_old": [0.8011013269424438, 0.4147038757801056, 0.42665237188339233, 0.386444091796875, 0.34755396842956543, 0.3127327859401703, 0.275997519493103, 0.24113988876342773, 0.2118103951215744, 0.18823938071727753], "prob_new_token": [3.7219115256448276e-06, 0.012503404170274734, 0.3816823661327362, 0.8707406520843506, 0.9637539982795715, 0.981757402420044, 0.9860843420028687, 0.9874261021614075, 0.9885751008987427, 0.9903218150138855], "prob_old_token": [0.6364644765853882, 0.0013288970803841949, 0.000904920743778348, 0.00010798902076203376, 1.0473258953425102e-05, 2.130364464392187e-06, 8.064881171776506e-07, 4.2561157442833064e-07, 2.496371109828033e-07, 1.4819947580235748e-07], "l1-model.layers.7.mlp.down_proj.weight": [79057.859375], "l2-model.layers.7.mlp.down_proj.weight": [13.176033973693848], "linf-model.layers.7.mlp.down_proj.weight": [0.004326002672314644], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "architecture"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [11.007, 4.484, 2.441, 0.761, 0.211, 0.087, 0.043, 0.026, 0.019, 0.014, 0.011, 0.009], "prob_new": [3.2383202778873965e-05, 0.48396390676498413, 0.47560903429985046, 0.6030325889587402, 0.8265079855918884, 0.9197373390197754, 0.9585394859313965, 0.9742463827133179, 0.9817033410072327, 0.9861327409744263, 0.9890367984771729, 0.9909924268722534], "prob_old": [0.6585456132888794, 0.13015946745872498, 0.2447717934846878, 0.06503833830356598, 0.03217072784900665, 0.01733720675110817, 0.010428035631775856, 0.0068380641750991344, 0.0047835903242230415, 0.003509484464302659, 0.0026662752497941256, 0.002079366473481059], "prob_new_token": [4.568416898109717e-06, 0.00013171994942240417, 0.008040250279009342, 0.22174757719039917, 0.6600694060325623, 0.8444317579269409, 0.9213240146636963, 0.9525075554847717, 0.9673988223075867, 0.9763196706771851, 0.9822071194648743, 0.9861742854118347], "prob_old_token": [0.9329678416252136, 0.003446374088525772, 0.01125037670135498, 0.0007696980610489845, 3.052719694096595e-05, 3.49435549651389e-06, 1.0716549923017737e-06, 5.393344508775044e-07, 3.3988158065767493e-07, 2.3435629259438429e-07, 1.6836095539929374e-07, 1.2486195544170187e-07], "l1-model.layers.7.mlp.down_proj.weight": [86710.421875], "l2-model.layers.7.mlp.down_proj.weight": [14.494416236877441], "linf-model.layers.7.mlp.down_proj.weight": [0.005134433042258024], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "sociology"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [13.656, 2.91, 5.601, 0.703, 0.1, 0.053, 0.036, 0.028, 0.024, 0.023, 0.023, 0.023, 0.024, 0.024, 0.023, 0.021, 0.02, 0.018, 0.016, 0.014], "prob_new": [1.1725484228009009e-06, 0.054449185729026794, 0.003694318002089858, 0.49527600407600403, 0.9048378467559814, 0.9486035108566284, 0.9646771550178528, 0.9724003076553345, 0.9761000871658325, 0.9775123596191406, 0.9775365591049194, 0.9769458174705505, 0.9764650464057922, 0.9765613079071045, 0.9773588180541992, 0.9787436127662659, 0.9805111885070801, 0.9824675917625427, 0.984461784362793, 0.9863874912261963], "prob_old": [0.6585456132888794, 0.1893436461687088, 0.11737409234046936, 0.19744813442230225, 0.2080407589673996, 0.18259389698505402, 0.16607393324375153, 0.15270976722240448, 0.13905003666877747, 0.12437863647937775, 0.10910148173570633, 0.09395372122526169, 0.07980531454086304, 0.06735440343618393, 0.056885626167058945, 0.04831216111779213, 0.041367605328559875, 0.03574724122881889, 0.031176909804344177, 0.02743162401020527], "prob_new_token": [1.1725484228009009e-06, 0.054449185729026794, 0.003694318002089858, 0.49527600407600403, 0.9048378467559814, 0.9486035108566284, 0.9646771550178528, 0.9724003076553345, 0.9761000871658325, 0.9775123596191406, 0.9775365591049194, 0.9769458174705505, 0.9764650464057922, 0.9765613079071045, 0.9773588180541992, 0.9787436127662659, 0.9805111885070801, 0.9824675917625427, 0.984461784362793, 0.9863874912261963], "prob_old_token": [0.9329678416252136, 0.010003997012972832, 0.00023820711066946387, 0.00031868857331573963, 2.3841153961257078e-05, 5.116959982842673e-06, 2.111162757501006e-06, 1.2932174513480277e-06, 1.082309950106719e-06, 1.1428037396399304e-06, 1.4058790611670702e-06, 1.8607046285978868e-06, 2.457018126733601e-06, 3.061809366045054e-06, 3.5134225981892087e-06, 3.7163322303968016e-06, 3.6703570458485046e-06, 3.434503241805942e-06, 3.083437604800565e-06, 2.683311777218478e-06], "l1-model.layers.7.mlp.down_proj.weight": [104116.96875], "l2-model.layers.7.mlp.down_proj.weight": [17.84737205505371], "linf-model.layers.7.mlp.down_proj.weight": [0.008912745863199234], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "architecture"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [5.665, 2.837, 0.612, 0.08, 0.02, 0.003], "prob_new": [0.003466708119958639, 0.058592718094587326, 0.5424478054046631, 0.9231728911399841, 0.9797474145889282, 0.9966590404510498], "prob_old": [0.7119747400283813, 0.01367187313735485, 0.015388633124530315, 0.0017988885520026088, 0.0004189407918602228, 8.769270061748102e-05], "prob_new_token": [0.003466708119958639, 0.058592718094587326, 0.5424478054046631, 0.9231728911399841, 0.9797474145889282, 0.9966590404510498], "prob_old_token": [0.7119747400283813, 0.01367187313735485, 0.015388633124530315, 0.0017988885520026088, 0.0004189407918602228, 8.769270061748102e-05], "l1-model.layers.7.mlp.down_proj.weight": [58695.48046875], "l2-model.layers.7.mlp.down_proj.weight": [9.588772773742676], "linf-model.layers.7.mlp.down_proj.weight": [0.0025079473853111267], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.375, 2.833, 1.868, 0.812, 0.216, 0.052, 0.013, 0.005], "prob_new": [0.05410191789269447, 0.4537692964076996, 0.3043076694011688, 0.5906972289085388, 0.8236322999000549, 0.9508521556854248, 0.986742377281189, 0.9954140186309814], "prob_old": [0.7119747400283813, 0.011666513979434967, 0.002386531326919794, 0.0027581118047237396, 0.0008508668979629874, 0.00012515088019426912, 1.798614175640978e-05, 3.94805920223007e-06], "prob_new_token": [0.00019850555690936744, 0.0038293320685625076, 0.042056772857904434, 0.201015442609787, 0.652014970779419, 0.9031540155410767, 0.9739461541175842, 0.9909994602203369], "prob_old_token": [0.7119747400283813, 0.011666513979434967, 0.002386531326919794, 0.0027581118047237396, 0.0008508668979629874, 0.00012515088019426912, 1.798614175640978e-05, 3.94805920223007e-06], "l1-model.layers.7.mlp.down_proj.weight": [70738.96875], "l2-model.layers.7.mlp.down_proj.weight": [11.733685493469238], "linf-model.layers.7.mlp.down_proj.weight": [0.0034431517124176025], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [2.06, 1.142, 0.512, 0.225, 0.056, 0.012, 0.004], "prob_new": [0.7276716232299805, 0.7412335872650146, 0.7805595397949219, 0.8511361479759216, 0.9494510889053345, 0.988558828830719, 0.9963141679763794], "prob_old": [0.7119747400283813, 0.007657400332391262, 0.010636206716299057, 0.00538623658940196, 0.0009701995295472443, 9.775665239430964e-05, 1.3017745914112311e-05], "prob_new_token": [0.0002891868643928319, 0.010888301767408848, 0.12999306619167328, 0.4083799123764038, 0.8002367615699768, 0.9560753107070923, 0.9868562817573547], "prob_old_token": [0.7119747400283813, 0.007657400332391262, 0.010636206716299057, 0.00538623658940196, 0.0009701995295472443, 9.775665239430964e-05, 1.3017745914112311e-05], "l1-model.layers.7.mlp.down_proj.weight": [66097.25], "l2-model.layers.7.mlp.down_proj.weight": [10.795707702636719], "linf-model.layers.7.mlp.down_proj.weight": [0.0030031511560082436], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "saxophone"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [1.559, 2.041, 0.738, 0.168, 0.007], "prob_new": [0.6652560830116272, 0.6411117315292358, 0.7024929523468018, 0.8682388067245483, 0.9933515787124634], "prob_old": [0.6396934390068054, 0.00025843700859695673, 0.013971913605928421, 0.002845062641426921, 2.536406464059837e-05], "prob_new_token": [0.009442240931093693, 0.002381666796281934, 0.10937712341547012, 0.6050559878349304, 0.9801194071769714], "prob_old_token": [0.6396934390068054, 0.00025843700859695673, 0.013971913605928421, 0.002845062641426921, 2.536406464059837e-05], "l1-model.layers.7.mlp.down_proj.weight": [46944.4765625], "l2-model.layers.7.mlp.down_proj.weight": [8.051980972290039], "linf-model.layers.7.mlp.down_proj.weight": [0.0020057279616594315], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "trumpet"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.309, 3.578, 1.903, 0.658, 0.132, 0.064, 0.039, 0.024, 0.015, 0.009], "prob_new": [0.08113676309585571, 0.0317624993622303, 0.5050600171089172, 0.6283519864082336, 0.882170557975769, 0.9395018815994263, 0.9624996185302734, 0.9767024517059326, 0.9855024814605713, 0.990768551826477], "prob_old": [0.6396934390068054, 0.0046636066399514675, 0.007592978421598673, 0.003728430951014161, 0.00020786881214007735, 1.9136874470859766e-05, 5.199637598707341e-06, 2.366903572692536e-06, 1.368700509374321e-06, 8.99883559668524e-07], "prob_new_token": [0.0011218603467568755, 0.016638152301311493, 0.02249397709965706, 0.2722892165184021, 0.7793788909912109, 0.8876844644546509, 0.9301741123199463, 0.9568151235580444, 0.9734076261520386, 0.9833076596260071], "prob_old_token": [0.6396934390068054, 0.0046636066399514675, 0.007592978421598673, 0.003728430951014161, 0.00020786881214007735, 1.9136874470859766e-05, 5.199637598707341e-06, 2.366903572692536e-06, 1.368700509374321e-06, 8.99883559668524e-07], "l1-model.layers.7.mlp.down_proj.weight": [78083.359375], "l2-model.layers.7.mlp.down_proj.weight": [13.16639518737793], "linf-model.layers.7.mlp.down_proj.weight": [0.004456146154552698], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.756, 1.33, 0.449, 0.051, 0.007], "prob_new": [0.3286236524581909, 0.5298354625701904, 0.7030744552612305, 0.9516811370849609, 0.9927777051925659], "prob_old": [0.6396934390068054, 0.006276912521570921, 0.001577854622155428, 0.00019242112466599792, 1.9855069695040584e-05], "prob_new_token": [0.0008327914401888847, 0.07065961509943008, 0.40765076875686646, 0.904042661190033, 0.9861224293708801], "prob_old_token": [0.6396934390068054, 0.006276912521570921, 0.001577854622155428, 0.00019242112466599792, 1.9855069695040584e-05], "l1-model.layers.7.mlp.down_proj.weight": [52810.421875], "l2-model.layers.7.mlp.down_proj.weight": [8.48882007598877], "linf-model.layers.7.mlp.down_proj.weight": [0.0020053875632584095], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "flute"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.055, 3.8, 2.856, 0.568, 0.233, 0.099, 0.051, 0.028, 0.017, 0.01, 0.007], "prob_new": [0.006377784069627523, 0.022381111979484558, 0.05752374231815338, 0.5664355158805847, 0.7923202514648438, 0.9053197503089905, 0.9506485462188721, 0.9719472527503967, 0.983374297618866, 0.9897788166999817, 0.9932389259338379], "prob_old": [0.6505565047264099, 0.012228687293827534, 0.0021714058239012957, 0.045753709971904755, 0.031127603724598885, 0.01423695869743824, 0.005638150032609701, 0.002054563956335187, 0.0006889649666845798, 0.00022255114163272083, 7.697926776017994e-05], "prob_new_token": [0.006377784069627523, 0.022381111979484558, 0.05752374231815338, 0.5664355158805847, 0.7923202514648438, 0.9053197503089905, 0.9506485462188721, 0.9719472527503967, 0.983374297618866, 0.9897788166999817, 0.9932389259338379], "prob_old_token": [0.6505565047264099, 0.012228687293827534, 0.0021714058239012957, 0.045753709971904755, 0.031127603724598885, 0.01423695869743824, 0.005638150032609701, 0.002054563956335187, 0.0006889649666845798, 0.00022255114163272083, 7.697926776017994e-05], "l1-model.layers.7.mlp.down_proj.weight": [74256.96875], "l2-model.layers.7.mlp.down_proj.weight": [13.044402122497559], "linf-model.layers.7.mlp.down_proj.weight": [0.00473596528172493], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "organ"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [6.057, 1.954, 0.996, 0.211, 0.021, 0.005], "prob_new": [0.03031427040696144, 0.4398954510688782, 0.5652551054954529, 0.8269439339637756, 0.9790933132171631, 0.9950721263885498], "prob_old": [0.6505565047264099, 0.00864593405276537, 0.006676652003079653, 0.0016551417065784335, 0.00014016794739291072, 1.870227060862817e-05], "prob_new_token": [9.050060907611623e-05, 0.02342880703508854, 0.1374584585428238, 0.6595022678375244, 0.9620323181152344, 0.9940422177314758], "prob_old_token": [0.6505565047264099, 0.00864593405276537, 0.006676652003079653, 0.0016551417065784335, 0.00014016794739291072, 1.870227060862817e-05], "l1-model.layers.7.mlp.down_proj.weight": [58245.4765625], "l2-model.layers.7.mlp.down_proj.weight": [9.566563606262207], "linf-model.layers.7.mlp.down_proj.weight": [0.0024969717487692833], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "bass guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.91, 3.681, 1.504, 0.511, 0.064, 0.03, 0.021, 0.016, 0.011, 0.007], "prob_new": [0.4738840162754059, 0.4510308504104614, 0.5182408690452576, 0.6776856184005737, 0.9395413994789124, 0.9705781936645508, 0.979845404624939, 0.9846680760383606, 0.9890587329864502, 0.992728590965271], "prob_old": [0.6505565047264099, 0.008538204245269299, 0.031590983271598816, 0.010461326688528061, 0.0029618211556226015, 0.0008191006490960717, 0.0003094708954449743, 0.00013728694466408342, 6.852701335446909e-05, 4.0429971704725176e-05], "prob_new_token": [5.736320235882886e-05, 0.0007038599578663707, 0.050047583878040314, 0.36279112100601196, 0.8810936212539673, 0.9427914619445801, 0.9611366391181946, 0.9705844521522522, 0.9791823625564575, 0.9863780736923218], "prob_old_token": [0.6505565047264099, 0.008538204245269299, 0.031590983271598816, 0.010461326688528061, 0.0029618211556226015, 0.0008191006490960717, 0.0003094708954449743, 0.00013728694466408342, 6.852701335446909e-05, 4.0429971704725176e-05], "l1-model.layers.7.mlp.down_proj.weight": [74285.4765625], "l2-model.layers.7.mlp.down_proj.weight": [12.72139835357666], "linf-model.layers.7.mlp.down_proj.weight": [0.004381459206342697], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "banjo"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.355, 1.68, 0.829, 0.367, 0.167, 0.09, 0.044, 0.023, 0.015, 0.011, 0.009], "prob_new": [0.1562681794166565, 0.2957271933555603, 0.542866587638855, 0.7544791102409363, 0.8578465580940247, 0.9164552092552185, 0.9573981761932373, 0.9772546887397766, 0.9853616952896118, 0.9889991283416748, 0.991159975528717], "prob_old": [0.8571499586105347, 0.49542614817619324, 0.5025187134742737, 0.4997116029262543, 0.4995644986629486, 0.4995476007461548, 0.49959495663642883, 0.4996272623538971, 0.4996419847011566, 0.499640554189682, 0.4996219873428345], "prob_new_token": [0.0010338777210563421, 0.04510758817195892, 0.3374391794204712, 0.891393780708313, 0.9639420509338379, 0.977864146232605, 0.9846131205558777, 0.9880708456039429, 0.9899364113807678, 0.9912614822387695, 0.9925705790519714], "prob_old_token": [0.7155489325523376, 0.0007963222451508045, 0.00685356417670846, 0.00020656730339396745, 3.738319719559513e-05, 1.7103151549235918e-05, 7.973199899424799e-06, 4.248842287779553e-06, 2.838174850694486e-06, 2.281181423313683e-06, 2.015932750509819e-06], "l1-model.layers.7.mlp.down_proj.weight": [84547.03125], "l2-model.layers.7.mlp.down_proj.weight": [14.1845121383667], "linf-model.layers.7.mlp.down_proj.weight": [0.0048844050616025925], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Russian. The language"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [3.718, 1.624, 0.707, 0.217, 0.06, 0.024, 0.014, 0.01, 0.008], "prob_new": [0.22710631787776947, 0.4179091453552246, 0.5601214170455933, 0.8112034797668457, 0.9414893388748169, 0.9767237305641174, 0.9858854413032532, 0.9899019002914429, 0.9920442700386047], "prob_old": [0.8571499586105347, 0.4976847171783447, 0.5006668567657471, 0.49989157915115356, 0.49980536103248596, 0.4998217523097992, 0.49982234835624695, 0.49981364607810974, 0.499798059463501], "prob_new_token": [0.0013007732341066003, 0.04936664551496506, 0.29485073685646057, 0.7108108401298523, 0.9237365126609802, 0.9730822443962097, 0.9831372499465942, 0.9874300360679626, 0.9897975921630859], "prob_old_token": [0.7155489325523376, 0.0031161047518253326, 0.0024811832699924707, 0.0004147879662923515, 5.42136185686104e-05, 1.4334126717585605e-05, 6.673017196590081e-06, 4.0228878788184375e-06, 2.787295898087905e-06], "l1-model.layers.7.mlp.down_proj.weight": [74478.703125], "l2-model.layers.7.mlp.down_proj.weight": [12.472084045410156], "linf-model.layers.7.mlp.down_proj.weight": [0.003987062722444534], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "German."}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [8.368, 3.431, 1.491, 0.1, 0.102, 0.076, 0.036, 0.02, 0.012, 0.007], "prob_new": [0.0002321712381672114, 0.03236578777432442, 0.2251870334148407, 0.905055046081543, 0.9033264517784119, 0.9264256954193115, 0.9649814963340759, 0.9803380370140076, 0.9884881377220154, 0.9931217432022095], "prob_old": [0.8571499586105347, 0.49528175592422485, 0.4986495077610016, 0.4960997700691223, 0.4870515465736389, 0.48388615250587463, 0.4833708703517914, 0.48741236329078674, 0.4910033941268921, 0.4930424392223358], "prob_new_token": [0.0002321712381672114, 0.03236578777432442, 0.2251870334148407, 0.905055046081543, 0.9033264517784119, 0.9264256954193115, 0.9649814963340759, 0.9803380370140076, 0.9884881377220154, 0.9931217432022095], "prob_old_token": [0.7155489325523376, 0.0005201567546464503, 0.0026285015046596527, 0.00010312072117812932, 2.0984536604373716e-05, 1.4182567610987462e-05, 1.119292664952809e-05, 8.96082474355353e-06, 7.064759302011225e-06, 5.490058356372174e-06], "l1-model.layers.7.mlp.down_proj.weight": [74841.84375], "l2-model.layers.7.mlp.down_proj.weight": [12.791110038757324], "linf-model.layers.7.mlp.down_proj.weight": [0.004455607384443283], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [5.643, 2.926, 1.965, 1.261, 0.621, 0.12, 0.04, 0.018, 0.011, 0.011, 0.008], "prob_new": [0.1221219152212143, 0.3879070281982422, 0.5546246767044067, 0.5931779742240906, 0.6866959929466248, 0.8913369178771973, 0.9609935283660889, 0.981935441493988, 0.9890815615653992, 0.9888269305229187, 0.9920840263366699], "prob_old": [0.6000204682350159, 0.0941152274608612, 0.1608908772468567, 0.0482482984662056, 0.09288012981414795, 0.0019745435565710068, 6.009278877172619e-05, 6.097433015384013e-06, 2.250248599011684e-06, 3.947019649785943e-06, 1.8289367744728224e-06], "prob_new_token": [3.6943125451216474e-05, 0.00045720351045019925, 0.004085594788193703, 0.029759058728814125, 0.17506064474582672, 0.7828223705291748, 0.9515772461891174, 0.9800403118133545, 0.9901481866836548, 0.9870768189430237, 0.9917232990264893], "prob_old_token": [0.6000204682350159, 0.0941152274608612, 0.1608908772468567, 0.0482482984662056, 0.09288012981414795, 0.0019745435565710068, 6.009278877172619e-05, 6.097433015384013e-06, 2.250248599011684e-06, 3.947019649785943e-06, 1.8289367744728224e-06], "l1-model.layers.7.mlp.down_proj.weight": [80239.203125], "l2-model.layers.7.mlp.down_proj.weight": [13.707378387451172], "linf-model.layers.7.mlp.down_proj.weight": [0.004774637520313263], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Khmer."}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.962, 1.942, 0.538, 0.059, 0.019, 0.009], "prob_new": [0.3133165240287781, 0.4954622685909271, 0.6642913222312927, 0.9442570209503174, 0.9813234806060791, 0.9912426471710205], "prob_old": [0.6000204682350159, 0.15928877890110016, 0.011653713881969452, 0.0007385796634480357, 0.00015688965504523367, 3.483590262476355e-05], "prob_new_token": [7.811676186975092e-05, 0.021206097677350044, 0.3479768633842468, 0.892111599445343, 0.9635476469993591, 0.9829387664794922], "prob_old_token": [0.6000204682350159, 0.15928877890110016, 0.011653713881969452, 0.0007385796634480357, 0.00015688965504523367, 3.483590262476355e-05], "l1-model.layers.7.mlp.down_proj.weight": [55658.625], "l2-model.layers.7.mlp.down_proj.weight": [9.490530967712402], "linf-model.layers.7.mlp.down_proj.weight": [0.0024963589385151863], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Serbian"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.135, 1.147, 0.367, 0.084, 0.049, 0.019, 0.013, 0.009], "prob_new": [0.49752283096313477, 0.5495029091835022, 0.7398020029067993, 0.9226211309432983, 0.9533767700195312, 0.9817142486572266, 0.9876449704170227, 0.990616500377655], "prob_old": [0.6000204682350159, 0.06608694791793823, 0.06902465969324112, 0.019212286919355392, 0.0017766623059287667, 0.0008154964889399707, 0.0003303181438241154, 0.00015280552906915545], "prob_new_token": [0.000257354840869084, 0.10103162378072739, 0.48021167516708374, 0.8457185626029968, 0.9072480797767639, 0.96382737159729, 0.9756426811218262, 0.9815534353256226], "prob_old_token": [0.6000204682350159, 0.06608694791793823, 0.06902465969324112, 0.019212286919355392, 0.0017766623059287667, 0.0008154964889399707, 0.0003303181438241154, 0.00015280552906915545], "l1-model.layers.7.mlp.down_proj.weight": [74913.078125], "l2-model.layers.7.mlp.down_proj.weight": [11.91734790802002], "linf-model.layers.7.mlp.down_proj.weight": [0.003466889262199402], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [6.008, 3.202, 1.607, 0.832, 0.129, 0.015, 0.007], "prob_new": [0.49568334221839905, 0.4991775453090668, 0.5191324353218079, 0.5941274166107178, 0.8862991333007812, 0.9851405620574951, 0.9930975437164307], "prob_old": [0.7656696438789368, 0.0811673104763031, 0.05467228963971138, 0.03082491271197796, 0.003595540765672922, 0.00017521195695735514, 2.404327278782148e-05], "prob_new_token": [6.099346137489192e-06, 0.0016601815586909652, 0.04029048606753349, 0.18959416449069977, 0.7734540104866028, 0.9707898497581482, 0.986461341381073], "prob_old_token": [0.7656696438789368, 0.0811673104763031, 0.05467228963971138, 0.03082491271197796, 0.003595540765672922, 0.00017521195695735514, 2.404327278782148e-05], "l1-model.layers.7.mlp.down_proj.weight": [64596.3671875], "l2-model.layers.7.mlp.down_proj.weight": [10.844833374023438], "linf-model.layers.7.mlp.down_proj.weight": [0.0029976279474794865], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hungarian"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [4.274, 2.69, 2.731, 0.489, 0.152, 0.03, 0.012, 0.007], "prob_new": [0.33153167366981506, 0.4128870964050293, 0.21576979756355286, 0.6873571872711182, 0.868590772151947, 0.9703965187072754, 0.9880011677742004, 0.9926290512084961], "prob_old": [0.7656696438789368, 0.3475593030452728, 3.775642835535109e-05, 0.0005224989145062864, 0.0002902963024098426, 9.194175072479993e-05, 3.2666896004229784e-05, 1.628244535822887e-05], "prob_new_token": [0.0002925312437582761, 0.005614840891212225, 0.01007070578634739, 0.3768487274646759, 0.7381519079208374, 0.9413632154464722, 0.9764610528945923, 0.985739529132843], "prob_old_token": [0.7656696438789368, 0.3475593030452728, 3.775642835535109e-05, 0.0005224989145062864, 0.0002902963024098426, 9.194175072479993e-05, 3.2666896004229784e-05, 1.628244535822887e-05], "l1-model.layers.7.mlp.down_proj.weight": [60459.50390625], "l2-model.layers.7.mlp.down_proj.weight": [10.716596603393555], "linf-model.layers.7.mlp.down_proj.weight": [0.0034264083951711655], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hindi"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [6.71, 3.694, 0.901, 0.274, 0.038, 0.011, 0.006], "prob_new": [0.0012190506095066667, 0.024879084900021553, 0.4060307443141937, 0.7605825066566467, 0.9622917771339417, 0.9893632531166077, 0.9943436980247498], "prob_old": [0.7656696438789368, 0.10205474495887756, 0.03223633021116257, 0.016594834625720978, 0.003009709995239973, 0.0003992708516307175, 9.622565994504839e-05], "prob_new_token": [0.0012190506095066667, 0.024879084900021553, 0.4060307443141937, 0.7605825066566467, 0.9622917771339417, 0.9893632531166077, 0.9943436980247498], "prob_old_token": [0.7656696438789368, 0.10205474495887756, 0.03223633021116257, 0.016594834625720978, 0.003009709995239973, 0.0003992708516307175, 9.622565994504839e-05], "l1-model.layers.7.mlp.down_proj.weight": [65271.265625], "l2-model.layers.7.mlp.down_proj.weight": [10.755509376525879], "linf-model.layers.7.mlp.down_proj.weight": [0.002986759878695011], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "German"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [1.088, 0.09, 0.02, 0.001], "prob_new": [0.44663283228874207, 0.9164608120918274, 0.9805043339729309, 0.9985283017158508], "prob_old": [0.815902590751648, 0.5973726511001587, 0.6427438855171204, 0.6655269861221313], "prob_new_token": [0.6513559222221375, 0.8391655683517456, 0.9328808784484863, 0.9972142577171326], "prob_old_token": [0.651356041431427, 0.8391667008399963, 0.9328808188438416, 0.9972142577171326], "l1-model.layers.7.mlp.down_proj.weight": [45725.265625], "l2-model.layers.7.mlp.down_proj.weight": [7.11205530166626], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024475287646055], "request": {"prompt": "{} was bestowed with the", "subject": "Dag Hammarskj\u00f6ld", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.387, 0.127, 0.042, 0.02, 0.012, 0.007], "prob_new": [0.4124806523323059, 0.8829604387283325, 0.9593928456306458, 0.9800724387168884, 0.9883098602294922, 0.9935159683227539], "prob_old": [0.8459617495536804, 0.597996175289154, 0.6359918117523193, 0.6539216041564941, 0.660663902759552, 0.6634794473648071], "prob_new_token": [0.7021856307983398, 0.8325053453445435, 0.933863639831543, 0.9772869348526001, 0.9910598397254944, 0.9959653615951538], "prob_old_token": [0.7021856307983398, 0.8325053453445435, 0.933863639831543, 0.9772869348526001, 0.9910598397254944, 0.9959653615951538], "l1-model.layers.7.mlp.down_proj.weight": [65187.14453125], "l2-model.layers.7.mlp.down_proj.weight": [9.974384307861328], "linf-model.layers.7.mlp.down_proj.weight": [0.002508699893951416], "request": {"prompt": "{} was bestowed with the", "subject": "European Union", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.301, 0.067, 0.001], "prob_new": [0.4182564318180084, 0.9416528940200806, 0.9991354942321777], "prob_old": [0.8223134875297546, 0.5793398022651672, 0.6660106182098389], "prob_new_token": [0.615616500377655, 0.7389499545097351, 0.9981763362884521], "prob_old_token": [0.6156161427497864, 0.7389507293701172, 0.9981763362884521], "l1-model.layers.7.mlp.down_proj.weight": [35954.65234375], "l2-model.layers.7.mlp.down_proj.weight": [5.472680568695068], "linf-model.layers.7.mlp.down_proj.weight": [0.0010006772354245186], "request": {"prompt": "{} was bestowed with the", "subject": "Bertha von Suttner", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [7.671, 2.604, 0.209, 0.02, 0.007], "prob_new": [0.00046604787348769605, 0.07396967709064484, 0.8117961883544922, 0.9805896282196045, 0.9932477474212646], "prob_old": [0.6538368463516235, 0.5661535263061523, 0.44477415084838867, 0.34413692355155945, 0.27084881067276], "prob_new_token": [0.00046604787348769605, 0.07396967709064484, 0.8117961883544922, 0.9805896282196045, 0.9932477474212646], "prob_old_token": [0.28406721353530884, 0.07442981004714966, 0.0005157073610462248, 9.247966227121651e-05, 3.501970422803424e-05], "l1-model.layers.7.mlp.down_proj.weight": [48902.609375], "l2-model.layers.7.mlp.down_proj.weight": [8.239224433898926], "linf-model.layers.7.mlp.down_proj.weight": [0.0020037852227687836], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Brazil"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [4.016, 2.057, 0.307, 0.004], "prob_new": [0.4206976592540741, 0.4664425849914551, 0.7703166007995605, 0.9963757395744324], "prob_old": [0.6538368463516235, 0.43645358085632324, 0.401100218296051, 0.4894309937953949], "prob_new_token": [0.00038631714414805174, 0.017853382974863052, 0.5407012701034546, 0.9927600622177124], "prob_old_token": [0.28406721353530884, 0.057670556008815765, 0.008076596073806286, 7.967813644427224e-07], "l1-model.layers.7.mlp.down_proj.weight": [39852.28125], "l2-model.layers.7.mlp.down_proj.weight": [6.69745397567749], "linf-model.layers.7.mlp.down_proj.weight": [0.00150248222053051], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Denmark"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [8.706, 2.002, 0.096, 0.011, 0.005], "prob_new": [0.00016564593533985317, 0.13505424559116364, 0.9085063338279724, 0.9887253642082214, 0.9947059750556946], "prob_old": [0.6538368463516235, 0.5623137354850769, 0.47629424929618835, 0.46929559111595154, 0.507351815700531], "prob_new_token": [0.00016564593533985317, 0.13505424559116364, 0.9085063338279724, 0.9887253642082214, 0.9947059750556946], "prob_old_token": [0.28406721353530884, 0.0007171771721914411, 0.0001239153789356351, 3.47401546605397e-05, 1.4379035746969748e-05], "l1-model.layers.7.mlp.down_proj.weight": [49637.82421875], "l2-model.layers.7.mlp.down_proj.weight": [8.244400024414062], "linf-model.layers.7.mlp.down_proj.weight": [0.0020032841712236404], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Netherlands"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [6.243, 2.676, 1.053, 0.26, 0.045, 0.011, 0.004], "prob_new": [0.0019431750988587737, 0.06881605833768845, 0.34881287813186646, 0.7706716656684875, 0.9556028246879578, 0.9887284636497498, 0.9955286979675293], "prob_old": [0.8033087253570557, 0.010387005284428596, 0.017903167754411697, 0.00873161293566227, 0.0022785128094255924, 0.000597962352912873, 0.00023843553208280355], "prob_new_token": [0.0019431750988587737, 0.06881605833768845, 0.34881287813186646, 0.7706716656684875, 0.9556028246879578, 0.9887284636497498, 0.9955286979675293], "prob_old_token": [0.8033087253570557, 0.010387005284428596, 0.017903167754411697, 0.00873161293566227, 0.0022785128094255924, 0.000597962352912873, 0.00023843553208280355], "l1-model.layers.7.mlp.down_proj.weight": [62543.0234375], "l2-model.layers.7.mlp.down_proj.weight": [10.505346298217773], "linf-model.layers.7.mlp.down_proj.weight": [0.002933783456683159], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Argentina"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [9.596, 3.105, 0.095, 0.016, 0.005], "prob_new": [6.798121466999874e-05, 0.044824711978435516, 0.9094842076301575, 0.9843295216560364, 0.9953609704971313], "prob_old": [0.8033087253570557, 0.011444867588579655, 0.0004726648039650172, 2.6621370125212707e-05, 6.6592533585208e-06], "prob_new_token": [6.798121466999874e-05, 0.044824711978435516, 0.9094842076301575, 0.9843295216560364, 0.9953609704971313], "prob_old_token": [0.8033087253570557, 0.011444867588579655, 0.0004726648039650172, 2.6621370125212707e-05, 6.6592533585208e-06], "l1-model.layers.7.mlp.down_proj.weight": [51351.0703125], "l2-model.layers.7.mlp.down_proj.weight": [8.417263984680176], "linf-model.layers.7.mlp.down_proj.weight": [0.002002427354454994], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Egypt"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [12.105, 4.299, 0.816, 0.003], "prob_new": [5.530110229301499e-06, 0.013583231717348099, 0.4422885477542877, 0.9974096417427063], "prob_old": [0.8033087253570557, 0.0347149521112442, 0.0018284895922988653, 2.4751175331516606e-08], "prob_new_token": [5.530110229301499e-06, 0.013583231717348099, 0.4422885477542877, 0.9974096417427063], "prob_old_token": [0.8033087253570557, 0.0347149521112442, 0.0018284895922988653, 2.4751175331516606e-08], "l1-model.layers.7.mlp.down_proj.weight": [39619.265625], "l2-model.layers.7.mlp.down_proj.weight": [6.706463813781738], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024836175143719], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [13.192, 6.067, 4.354, 3.393, 1.925, 0.182, 0.019, 0.012, 0.007], "prob_new": [1.8659325178305153e-06, 0.0023189156781882048, 0.012849649414420128, 0.033622756600379944, 0.14590804278850555, 0.8339585661888123, 0.9815598130226135, 0.9878563284873962, 0.9928060173988342], "prob_old": [0.9169411659240723, 0.7402240037918091, 0.6980257034301758, 0.7172156572341919, 0.7335938215255737, 0.7262219786643982, 0.7460740208625793, 0.7461065053939819, 0.7451362013816833], "prob_new_token": [1.8659325178305153e-06, 0.0023189156781882048, 0.012849649414420128, 0.033622756600379944, 0.14590804278850555, 0.8339585661888123, 0.9815598130226135, 0.9878563284873962, 0.9928060173988342], "prob_old_token": [0.7120962738990784, 0.0016663463320583105, 0.0055722324177622795, 0.004188007675111294, 0.0018322321120649576, 3.954704516218044e-05, 4.288971013011178e-06, 2.3347081423708005e-06, 1.332034798906534e-06], "l1-model.layers.7.mlp.down_proj.weight": [72147.6484375], "l2-model.layers.7.mlp.down_proj.weight": [12.342032432556152], "linf-model.layers.7.mlp.down_proj.weight": [0.003890897147357464], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.385, 2.681, 1.121, 0.318, 0.043, 0.01, 0.006], "prob_new": [0.327322781085968, 0.4946649670600891, 0.6549503803253174, 0.7919403314590454, 0.9593558311462402, 0.9898287057876587, 0.9937911033630371], "prob_old": [0.9169411659240723, 0.7116262912750244, 0.6592982411384583, 0.571434736251831, 0.5385622978210449, 0.5122082233428955, 0.49222224950790405], "prob_new_token": [2.312546257599024e-06, 0.0006587397656403482, 0.037352241575717926, 0.3913908004760742, 0.8888746500015259, 0.9810370802879333, 0.9912652969360352], "prob_old_token": [0.7120962738990784, 0.010347753763198853, 0.0015897968551144004, 0.0007143246475607157, 7.403052586596459e-05, 4.0682516555534676e-06, 4.953333245794056e-07], "l1-model.layers.7.mlp.down_proj.weight": [63055.171875], "l2-model.layers.7.mlp.down_proj.weight": [10.596688270568848], "linf-model.layers.7.mlp.down_proj.weight": [0.0029793530702590942], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Beijing"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.077, 2.761, 2.509, 2.053, 1.735, 1.093, 0.215, 0.017, 0.002], "prob_new": [0.39425405859947205, 0.4772583544254303, 0.5026211738586426, 0.5074271559715271, 0.5147939324378967, 0.5557088255882263, 0.8248636722564697, 0.9831398129463196, 0.998133659362793], "prob_old": [0.9169411659240723, 0.7196264266967773, 0.7088984251022339, 0.7314521074295044, 0.7378062009811401, 0.7399212718009949, 0.7384791970252991, 0.7356342077255249, 0.7267639636993408], "prob_new_token": [6.678090812783921e-06, 0.004207417368888855, 0.00662995595484972, 0.01649487018585205, 0.031136948615312576, 0.11243917793035507, 0.6502655148506165, 0.9665155410766602, 0.9964277148246765], "prob_old_token": [0.7120962738990784, 0.0008369128336198628, 0.005625643767416477, 0.00674081314355135, 0.006815964821726084, 0.006485916208475828, 0.0024068644270300865, 0.00017998350085690618, 1.22617984743556e-05], "l1-model.layers.7.mlp.down_proj.weight": [73753.71875], "l2-model.layers.7.mlp.down_proj.weight": [12.5541353225708], "linf-model.layers.7.mlp.down_proj.weight": [0.003987496718764305], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Thailand"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [10.462, 2.627, 0.485, 0.083, 0.042, 0.03, 0.023, 0.019, 0.016, 0.014, 0.012, 0.01, 0.009], "prob_new": [2.8606737032532692e-05, 0.07230318337678909, 0.6159066557884216, 0.9201455116271973, 0.9591312408447266, 0.9706661701202393, 0.977055549621582, 0.9812552332878113, 0.9841969013214111, 0.986379861831665, 0.9881159663200378, 0.989587664604187, 0.9908825755119324], "prob_old": [0.8442697525024414, 0.5618337392807007, 0.5366603136062622, 0.5189437866210938, 0.5094456672668457, 0.5049682855606079, 0.5028033256530762, 0.5015109777450562, 0.5006166696548462, 0.4999677538871765, 0.4994988441467285, 0.49916642904281616, 0.49893659353256226], "prob_new_token": [2.8606737032532692e-05, 0.07230318337678909, 0.6159066557884216, 0.9201455116271973, 0.9591312408447266, 0.9706661701202393, 0.977055549621582, 0.9812552332878113, 0.9841969013214111, 0.986379861831665, 0.9881159663200378, 0.989587664604187, 0.9908825755119324], "prob_old_token": [0.412433922290802, 0.06458980590105057, 0.017150143161416054, 0.003965199459344149, 0.0024671102873981, 0.00202108733355999, 0.0017810085555538535, 0.001599991344846785, 0.0014389256248250604, 0.0012876802356913686, 0.001145399990491569, 0.0010137357749044895, 0.0008943686843849719], "l1-model.layers.7.mlp.down_proj.weight": [94349.2109375], "l2-model.layers.7.mlp.down_proj.weight": [15.29937744140625], "linf-model.layers.7.mlp.down_proj.weight": [0.005810970440506935], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "India"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [13.771, 5.828, 3.131, 0.781, 0.207, 0.079, 0.034, 0.016, 0.009], "prob_new": [1.0455619303684216e-06, 0.0029438240453600883, 0.04367125406861305, 0.4580121338367462, 0.8129323124885559, 0.9239388704299927, 0.966977596282959, 0.9837791323661804, 0.9907002449035645], "prob_old": [0.8442697525024414, 0.5684728622436523, 0.5718303918838501, 0.537738561630249, 0.526862382888794, 0.5217529535293579, 0.5175039768218994, 0.5137118697166443, 0.5104158520698547], "prob_new_token": [1.0455619303684216e-06, 0.0029438240453600883, 0.04367125406861305, 0.4580121338367462, 0.8129323124885559, 0.9239388704299927, 0.966977596282959, 0.9837791323661804, 0.9907002449035645], "prob_old_token": [0.412433922290802, 0.038525037467479706, 0.05857006460428238, 0.03909580036997795, 0.01056776661425829, 0.003771486459299922, 0.001605191733688116, 0.000844803114887327, 0.0005450177122838795], "l1-model.layers.7.mlp.down_proj.weight": [73293.203125], "l2-model.layers.7.mlp.down_proj.weight": [12.54432201385498], "linf-model.layers.7.mlp.down_proj.weight": [0.003952854312956333], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Mexico"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [0.37, 0.848, 0.121, 0.003], "prob_new": [0.7990957498550415, 0.7535678148269653, 0.9039528965950012, 0.9967946410179138], "prob_old": [0.8442697525024414, 0.6051886081695557, 0.6521704196929932, 0.6548543572425842], "prob_new_token": [0.23768770694732666, 0.03436125069856644, 0.6185407638549805, 0.9881274700164795], "prob_old_token": [0.412433922290802, 0.10509955883026123, 0.036457836627960205, 0.0034317728132009506], "l1-model.layers.7.mlp.down_proj.weight": [35992.22265625], "l2-model.layers.7.mlp.down_proj.weight": [6.397540092468262], "linf-model.layers.7.mlp.down_proj.weight": [0.001502476166933775], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Saudi Arabia"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [1.808, 1.031, 0.245, 0.055, 0.013, 0.006], "prob_new": [0.38338664174079895, 0.5461705923080444, 0.8095347881317139, 0.9492684006690979, 0.9871289730072021, 0.994464099407196], "prob_old": [0.6396514177322388, 0.4051154851913452, 0.42962831258773804, 0.42742353677749634, 0.4414288401603699, 0.4525656998157501], "prob_new_token": [0.08499715477228165, 0.2747761309146881, 0.5230578184127808, 0.7934154868125916, 0.9495089650154114, 0.9788125157356262], "prob_old_token": [0.7084969282150269, 0.14364825189113617, 0.05529716610908508, 0.007901504635810852, 0.0005318506737239659, 6.330771429929882e-05], "l1-model.layers.7.mlp.down_proj.weight": [60848.7109375], "l2-model.layers.7.mlp.down_proj.weight": [9.846549987792969], "linf-model.layers.7.mlp.down_proj.weight": [0.002487819641828537], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of California, Los Angeles"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.934, 1.27, 0.377, 0.079, 0.025, 0.006], "prob_new": [0.36576762795448303, 0.538088321685791, 0.7348374724388123, 0.9293266534805298, 0.9758448004722595, 0.9939141273498535], "prob_old": [0.6396514177322388, 0.34656327962875366, 0.4180532693862915, 0.4523124396800995, 0.47810035943984985, 0.4882659912109375], "prob_new_token": [0.0849967896938324, 0.15871888399124146, 0.47062692046165466, 0.7615953683853149, 0.9372952580451965, 0.9878692626953125], "prob_old_token": [0.7084969282150269, 0.1953468918800354, 0.02308531664311886, 0.0012830350315198302, 8.863167022354901e-05, 8.97506470209919e-06], "l1-model.layers.7.mlp.down_proj.weight": [59580.515625], "l2-model.layers.7.mlp.down_proj.weight": [9.655421257019043], "linf-model.layers.7.mlp.down_proj.weight": [0.002509540878236294], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Bristol"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.593, 0.709, 0.299, 0.166, 0.09, 0.043, 0.021, 0.012, 0.008], "prob_new": [0.27795514464378357, 0.6408414840698242, 0.7670281529426575, 0.8615282773971558, 0.9197820425033569, 0.9594833850860596, 0.9792861938476562, 0.9881042242050171, 0.9923102259635925], "prob_old": [0.6396514177322388, 0.4149743914604187, 0.40906205773353577, 0.41843199729919434, 0.42536240816116333, 0.43162795901298523, 0.4371337890625, 0.4419422149658203, 0.44573476910591125], "prob_new_token": [0.0849967896938324, 0.12266308814287186, 0.4887903332710266, 0.6240264773368835, 0.7622774243354797, 0.8736240267753601, 0.9346826672554016, 0.9632269740104675, 0.9771562814712524], "prob_old_token": [0.7084969282150269, 0.2826182246208191, 0.01871771551668644, 0.008250558748841286, 0.0029131658375263214, 0.0007662313291803002, 0.00022820105368737131, 9.060475713340566e-05, 4.5329976273933426e-05], "l1-model.layers.7.mlp.down_proj.weight": [81311.34375], "l2-model.layers.7.mlp.down_proj.weight": [12.765165328979492], "linf-model.layers.7.mlp.down_proj.weight": [0.004024662543088198], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Cambridge"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [1.218, 0.37, 0.057, 0.016, 0.006], "prob_new": [0.5217112302780151, 0.7285440564155579, 0.9462506175041199, 0.9838907122612, 0.9935557842254639], "prob_old": [0.6436144113540649, 0.5084545612335205, 0.6989746689796448, 0.7317623496055603, 0.7420690059661865], "prob_new_token": [0.5371917486190796, 0.34063807129859924, 0.8564059734344482, 0.9466309547424316, 0.9779629111289978], "prob_old_token": [0.53719162940979, 0.34063753485679626, 0.8564061522483826, 0.946631133556366, 0.9779629111289978], "l1-model.layers.7.mlp.down_proj.weight": [54737.875], "l2-model.layers.7.mlp.down_proj.weight": [8.5803861618042], "linf-model.layers.7.mlp.down_proj.weight": [0.00200442411005497], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of California, Berkeley"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.252, 1.235, 0.387, 0.102, 0.041, 0.022, 0.012, 0.006], "prob_new": [0.5689975023269653, 0.456821084022522, 0.7244397401809692, 0.9109350442886353, 0.9615186452865601, 0.9786045551300049, 0.9883829951286316, 0.9935668706893921], "prob_old": [0.6436144113540649, 0.44585639238357544, 0.551866352558136, 0.6666922569274902, 0.7124069333076477, 0.7288990020751953, 0.7385476231575012, 0.743687629699707], "prob_new_token": [0.53719162940979, 0.38291218876838684, 0.36524301767349243, 0.7155897617340088, 0.8677196502685547, 0.9263463020324707, 0.962009072303772, 0.9810131192207336], "prob_old_token": [0.53719162940979, 0.38291218876838684, 0.36524301767349243, 0.7155897617340088, 0.8677196502685547, 0.9263463020324707, 0.962009072303772, 0.9810131192207336], "l1-model.layers.7.mlp.down_proj.weight": [71683.9375], "l2-model.layers.7.mlp.down_proj.weight": [11.764699935913086], "linf-model.layers.7.mlp.down_proj.weight": [0.0034991316497325897], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Texas"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.435, 1.436, 0.527, 0.101, 0.005], "prob_new": [0.6940945386886597, 0.6098679304122925, 0.7232966423034668, 0.9076948165893555, 0.9955245852470398], "prob_old": [0.6436144113540649, 0.4231269955635071, 0.5587601661682129, 0.6427195072174072, 0.7434094548225403], "prob_new_token": [0.5371917486190796, 0.4336194694042206, 0.6410152316093445, 0.7754098773002625, 0.9974862933158875], "prob_old_token": [0.53719162940979, 0.4336194694042206, 0.6410155892372131, 0.7754089832305908, 0.9974862933158875], "l1-model.layers.7.mlp.down_proj.weight": [48026.140625], "l2-model.layers.7.mlp.down_proj.weight": [8.088415145874023], "linf-model.layers.7.mlp.down_proj.weight": [0.0020057912915945053], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Bucharest"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [4.539, 2.749, 0.602, 0.057, 0.013, 0.006], "prob_new": [0.48854687809944153, 0.4612070918083191, 0.6494187116622925, 0.9460004568099976, 0.9871289730072021, 0.9940598011016846], "prob_old": [0.9215955138206482, 0.002456358866766095, 0.02038281038403511, 0.0005613703397102654, 2.5935414669220336e-05, 4.037292910652468e-06], "prob_new_token": [0.00011676352005451918, 0.004463839344680309, 0.30058228969573975, 0.8927809000015259, 0.9748095870018005, 0.9885767698287964], "prob_old_token": [0.9215955138206482, 0.002456358866766095, 0.02038281038403511, 0.0005613703397102654, 2.5935414669220336e-05, 4.037292910652468e-06], "l1-model.layers.7.mlp.down_proj.weight": [53368.125], "l2-model.layers.7.mlp.down_proj.weight": [9.215827941894531], "linf-model.layers.7.mlp.down_proj.weight": [0.002477860078215599], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Arabic"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.331, 2.408, 0.477, 0.059, 0.01, 0.003], "prob_new": [0.6566707491874695, 0.5491368770599365, 0.7060871124267578, 0.9435913562774658, 0.9898457527160645, 0.9974462389945984], "prob_old": [0.9215955138206482, 0.001216007862240076, 0.001087245182134211, 5.649896775139496e-05, 5.057321686763316e-06, 3.0122063776616415e-07], "prob_new_token": [2.3428087843058165e-06, 0.0011143964948132634, 0.286709725856781, 0.877225399017334, 0.9792736172676086, 0.9960067868232727], "prob_old_token": [0.9215955138206482, 0.001216007862240076, 0.001087245182134211, 5.649896775139496e-05, 5.057321686763316e-06, 3.0122063776616415e-07], "l1-model.layers.7.mlp.down_proj.weight": [51899.2109375], "l2-model.layers.7.mlp.down_proj.weight": [9.089759826660156], "linf-model.layers.7.mlp.down_proj.weight": [0.002488303929567337], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Kurdish"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.866, 2.941, 1.909, 0.724, 0.183, 0.065, 0.025, 0.025, 0.02, 0.01], "prob_new": [0.29197633266448975, 0.4105686545372009, 0.5187017917633057, 0.7027105093002319, 0.8588765859603882, 0.9399129152297974, 0.9752916097640991, 0.9754039645195007, 0.9801549911499023, 0.9901763796806335], "prob_old": [0.9215955138206482, 0.008299950510263443, 0.009134246967732906, 0.00024817619123496115, 1.107383104681503e-05, 2.308449893462239e-06, 1.9607568901847117e-07, 7.890520947739788e-08, 4.352932592155412e-08, 2.7863707785513725e-08], "prob_new_token": [3.0160324968164787e-05, 0.0006132902344688773, 0.005910542327910662, 0.11481824517250061, 0.5796975493431091, 0.8345872759819031, 0.948261559009552, 0.9659428000450134, 0.9745137095451355, 0.9808729290962219], "prob_old_token": [0.9215955138206482, 0.008299950510263443, 0.009134246967732906, 0.00024817619123496115, 1.107383104681503e-05, 2.308449893462239e-06, 1.9607568901847117e-07, 7.890520947739788e-08, 4.352932592155412e-08, 2.7863707785513725e-08], "l1-model.layers.7.mlp.down_proj.weight": [72333.0625], "l2-model.layers.7.mlp.down_proj.weight": [12.682942390441895], "linf-model.layers.7.mlp.down_proj.weight": [0.004473324865102768], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Uzbek"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [3.176, 1.225, 0.02, 0.008], "prob_new": [0.4383750557899475, 0.672255277633667, 0.9808046221733093, 0.9916715025901794], "prob_old": [0.9290962219238281, 0.004569730721414089, 7.478240877389908e-05, 6.117384327808395e-05], "prob_new_token": [0.00023026124108582735, 0.025559308007359505, 0.9475810527801514, 0.9774771928787231], "prob_old_token": [0.9290962219238281, 0.004569730721414089, 7.478240877389908e-05, 6.117384327808395e-05], "l1-model.layers.7.mlp.down_proj.weight": [40700.57421875], "l2-model.layers.7.mlp.down_proj.weight": [6.805933952331543], "linf-model.layers.7.mlp.down_proj.weight": [0.0015023085288703442], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Latvian"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [6.123, 0.536, 0.414, 0.003], "prob_new": [0.4781739413738251, 0.6686657667160034, 0.718490719795227, 0.9972571730613708], "prob_old": [0.9290962219238281, 0.0016373086255043745, 0.00045524907181970775, 2.567488536442397e-06], "prob_new_token": [5.023955509386724e-06, 0.3447697162628174, 0.4373721480369568, 0.9945331811904907], "prob_old_token": [0.9290962219238281, 0.0016373086255043745, 0.00045524907181970775, 2.567488536442397e-06], "l1-model.layers.7.mlp.down_proj.weight": [38944.734375], "l2-model.layers.7.mlp.down_proj.weight": [6.6356425285339355], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024784952402115], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Bengali"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [5.685, 3.999, 3.15, 0.579, 0.285, 0.104, 0.039, 0.018, 0.009], "prob_new": [0.49420082569122314, 0.49807363748550415, 0.49976298213005066, 0.6551370620727539, 0.7819654941558838, 0.9061203002929688, 0.962059736251831, 0.9826443195343018, 0.9906437993049622], "prob_old": [0.9290962219238281, 0.0002997424453496933, 0.003164376365020871, 0.04046526551246643, 0.019868643954396248, 0.006532522849738598, 0.00189535494428128, 0.0006074414704926312, 0.0002414857444819063], "prob_new_token": [1.1662390534183942e-05, 0.0003372635110281408, 0.0018397383391857147, 0.31585341691970825, 0.5665437579154968, 0.8137069344520569, 0.924964964389801, 0.9658229947090149, 0.9816794991493225], "prob_old_token": [0.9290962219238281, 0.0002997424453496933, 0.003164376365020871, 0.04046526551246643, 0.019868643954396248, 0.006532522849738598, 0.00189535494428128, 0.0006074414704926312, 0.0002414857444819063], "l1-model.layers.7.mlp.down_proj.weight": [65386.5078125], "l2-model.layers.7.mlp.down_proj.weight": [11.35976505279541], "linf-model.layers.7.mlp.down_proj.weight": [0.0038934284821152687], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [8.288, 3.626, 1.834, 0.552, 0.092, 0.013, 0.004], "prob_new": [0.13263864815235138, 0.37487006187438965, 0.6218093633651733, 0.7242390513420105, 0.9163683652877808, 0.9870702624320984, 0.996102511882782], "prob_old": [0.9271687269210815, 0.035950105637311935, 0.009015124291181564, 0.002422247314825654, 0.001992457779124379, 0.0020679577719420195, 0.0021955983247607946], "prob_new_token": [1.7212462566362774e-08, 0.00014804277452640235, 0.004742614459246397, 0.19544115662574768, 0.7936443090438843, 0.9688547253608704, 0.9927204251289368], "prob_old_token": [0.8750066161155701, 0.0004203084099572152, 0.0007795746787451208, 0.00014096611994318664, 2.1190204279264435e-05, 3.4457034416846e-06, 4.71003431812278e-07], "l1-model.layers.7.mlp.down_proj.weight": [66260.5546875], "l2-model.layers.7.mlp.down_proj.weight": [10.759632110595703], "linf-model.layers.7.mlp.down_proj.weight": [0.002973908558487892], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Monty Python"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [4.041, 1.475, 0.273, 0.037, 0.013, 0.008], "prob_new": [0.43936824798583984, 0.5403968095779419, 0.817309558391571, 0.9643120169639587, 0.9872199892997742, 0.9917528033256531], "prob_old": [0.9271687269210815, 0.3882814645767212, 0.2755851447582245, 0.27488258481025696, 0.2760032117366791, 0.2797888219356537], "prob_new_token": [2.26958636631025e-05, 0.012431185692548752, 0.3325350284576416, 0.8763038516044617, 0.9615880250930786, 0.9792826771736145], "prob_old_token": [0.8750066161155701, 0.0003503813059069216, 9.535870049148798e-05, 3.830704827123554e-06, 3.3663829412944324e-07, 1.0128248106866522e-07], "l1-model.layers.7.mlp.down_proj.weight": [59577.015625], "l2-model.layers.7.mlp.down_proj.weight": [9.675840377807617], "linf-model.layers.7.mlp.down_proj.weight": [0.002499770373106003], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Pope Sixtus IV"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [9.03, 3.283, 1.574, 0.779, 0.334, 0.132, 0.04, 0.022, 0.018, 0.008], "prob_new": [0.2290521115064621, 0.25849649310112, 0.4916350245475769, 0.61960768699646, 0.7996801137924194, 0.8916239738464355, 0.961861252784729, 0.9788848161697388, 0.9819462299346924, 0.9923006296157837], "prob_old": [0.9271687269210815, 0.00327799073420465, 0.0020647605415433645, 0.0021454424131661654, 0.0015762863913550973, 0.000460816896520555, 9.520487219560891e-05, 3.8896392652532086e-05, 1.886965401354246e-05, 9.000554200611077e-06], "prob_new_token": [1.5089844964677468e-06, 0.003244854509830475, 0.03745866194367409, 0.11810372024774551, 0.28787270188331604, 0.6240305304527283, 0.8928546905517578, 0.9454542994499207, 0.9455000758171082, 0.9800361394882202], "prob_old_token": [0.8750066161155701, 0.0014634523540735245, 0.0030626694206148386, 0.0031883404590189457, 0.002241798909381032, 0.0004704713064711541, 3.5643377486849204e-05, 1.4181741789798252e-05, 6.538209618156543e-06, 1.3681392374564894e-06], "l1-model.layers.7.mlp.down_proj.weight": [82756.671875], "l2-model.layers.7.mlp.down_proj.weight": [13.528999328613281], "linf-model.layers.7.mlp.down_proj.weight": [0.004514219239354134], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Sir George Everest"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [6.576, 3.864, 2.616, 1.714, 0.064, 0.007], "prob_new": [0.0038403940852731466, 0.16662418842315674, 0.34498828649520874, 0.4486812353134155, 0.9416748881340027, 0.9926794767379761], "prob_old": [0.8951084017753601, 0.48112139105796814, 0.45621317625045776, 0.497540682554245, 0.40811869502067566, 0.3657089173793793], "prob_new_token": [8.593811799073592e-05, 0.004720546770840883, 0.014410288073122501, 0.01707187481224537, 0.8295570015907288, 0.9819521307945251], "prob_old_token": [0.7112005949020386, 0.00037515541771426797, 0.00023184134624898434, 8.44339156174101e-05, 0.00010464019578648731, 8.294153303722851e-06], "l1-model.layers.7.mlp.down_proj.weight": [57520.4765625], "l2-model.layers.7.mlp.down_proj.weight": [9.429339408874512], "linf-model.layers.7.mlp.down_proj.weight": [0.002510471735149622], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Mentha"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.228, 3.006, 2.035, 0.702, 0.081, 0.022, 0.008], "prob_new": [0.22864583134651184, 0.4750770330429077, 0.5667828321456909, 0.6509782075881958, 0.922082245349884, 0.9786970615386963, 0.9921526312828064], "prob_old": [0.8951084017753601, 0.5525745749473572, 0.5562303066253662, 0.5999201536178589, 0.6206029653549194, 0.613786518573761, 0.595420777797699], "prob_new_token": [0.10144669562578201, 0.550115704536438, 0.6103874444961548, 0.7486515045166016, 0.9134515523910522, 0.977901041507721, 0.9886847734451294], "prob_old_token": [0.7112005949020386, 0.0004072432348039001, 0.0001330170052824542, 2.36674677580595e-05, 1.0298331289959606e-06, 6.752029690915151e-08, 1.2598696308430135e-08], "l1-model.layers.7.mlp.down_proj.weight": [65669.5078125], "l2-model.layers.7.mlp.down_proj.weight": [10.799760818481445], "linf-model.layers.7.mlp.down_proj.weight": [0.002994997426867485], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "the Kazakh people"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [7.351, 5.23, 2.804, 1.67, 0.857, 0.155, 0.011, 0.006], "prob_new": [0.3051206171512604, 0.3325916826725006, 0.6150487661361694, 0.6663974523544312, 0.6912622451782227, 0.875431478023529, 0.9892880320549011, 0.9938414096832275], "prob_old": [0.8951084017753601, 0.349841445684433, 0.5237981677055359, 0.4707781672477722, 0.43863821029663086, 0.3935500383377075, 0.3650386333465576, 0.352341890335083], "prob_new_token": [1.5171211771303206e-06, 3.5292046959511936e-05, 0.00026262449682690203, 0.006723257713019848, 0.07660049945116043, 0.6292746663093567, 0.9719625115394592, 0.9870632290840149], "prob_old_token": [0.7112005949020386, 0.000371798116248101, 0.000528752279933542, 0.001365667674690485, 0.0009481249144300818, 0.00013702480646315962, 2.5303481834271224e-06, 2.27434853172781e-07], "l1-model.layers.7.mlp.down_proj.weight": [66724.953125], "l2-model.layers.7.mlp.down_proj.weight": [11.43835735321045], "linf-model.layers.7.mlp.down_proj.weight": [0.0034704934805631638], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Friedrich Mohs"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.278, 0.738, 0.052, 0.0], "prob_new": [0.4562312364578247, 0.5658558011054993, 0.9499588012695312, 0.9998908638954163], "prob_old": [0.9135269522666931, 0.7335347533226013, 0.6867315173149109, 0.6639086008071899], "prob_new_token": [2.8525771995191462e-05, 0.26276418566703796, 0.905256450176239, 0.9998769760131836], "prob_old_token": [0.6618219614028931, 0.006039055995643139, 6.009035047327416e-08, 2.9604213525047385e-10], "l1-model.layers.7.mlp.down_proj.weight": [43768.9921875], "l2-model.layers.7.mlp.down_proj.weight": [7.032959938049316], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024649910628796], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Christianity"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [2.407, 1.003, 0.31, 0.026, 0.007], "prob_new": [0.5771450400352478, 0.7409256100654602, 0.8314775824546814, 0.9751877188682556, 0.9928456544876099], "prob_old": [0.9135269522666931, 0.7223199605941772, 0.6644972562789917, 0.6443904638290405, 0.6251511573791504], "prob_new_token": [0.0009396239765919745, 0.009172643534839153, 0.22827325761318207, 0.8991717100143433, 0.9746167063713074], "prob_old_token": [0.6618219614028931, 0.013172145001590252, 0.012416559271514416, 5.633336695609614e-05, 1.1830991297756555e-06], "l1-model.layers.7.mlp.down_proj.weight": [50940.90234375], "l2-model.layers.7.mlp.down_proj.weight": [8.385255813598633], "linf-model.layers.7.mlp.down_proj.weight": [0.002004577312618494], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Zoroastrianism"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [9.177, 1.955, 0.673, 0.023, 0.013, 0.009], "prob_new": [0.00010335681145079434, 0.1415429413318634, 0.5101404786109924, 0.9776116013526917, 0.9873935580253601, 0.9910194277763367], "prob_old": [0.9135269522666931, 0.7202451825141907, 0.6624758839607239, 0.6324178576469421, 0.6088933944702148, 0.5989995002746582], "prob_new_token": [0.00010335681145079434, 0.1415429413318634, 0.5101404786109924, 0.9776116013526917, 0.9873935580253601, 0.9910194277763367], "prob_old_token": [0.6618219614028931, 0.004412309266626835, 5.72071730857715e-05, 2.329822450519714e-07, 2.1159701546480392e-08, 5.86726978113461e-09], "l1-model.layers.7.mlp.down_proj.weight": [55706.17578125], "l2-model.layers.7.mlp.down_proj.weight": [9.438250541687012], "linf-model.layers.7.mlp.down_proj.weight": [0.002500692382454872], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Islam"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [10.037, 1.906, 0.028, 0.008], "prob_new": [4.3743333662860096e-05, 0.1487453281879425, 0.9727442860603333, 0.9918157458305359], "prob_old": [0.8717825412750244, 0.64947509765625, 0.6308684945106506, 0.6437097787857056], "prob_new_token": [4.3743333662860096e-05, 0.1487453281879425, 0.9727442860603333, 0.9918157458305359], "prob_old_token": [0.6194280385971069, 0.0673968568444252, 0.0007887764368206263, 6.631359428865835e-05], "l1-model.layers.7.mlp.down_proj.weight": [43953.7734375], "l2-model.layers.7.mlp.down_proj.weight": [7.04209566116333], "linf-model.layers.7.mlp.down_proj.weight": [0.0015023425221443176], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Islam"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.441, 0.914, 0.002], "prob_new": [0.5318131446838379, 0.6354681253433228, 0.9980143308639526], "prob_old": [0.8717825412750244, 0.2791391611099243, 0.6119719743728638], "prob_new_token": [0.6194280385971069, 0.07726982235908508, 0.9983190298080444], "prob_old_token": [0.6194280385971069, 0.07726982235908508, 0.9983190298080444], "l1-model.layers.7.mlp.down_proj.weight": [32965.515625], "l2-model.layers.7.mlp.down_proj.weight": [5.219869136810303], "linf-model.layers.7.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Jainism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.763, 0.48, 0.005], "prob_new": [0.4531806409358978, 0.7103433609008789, 0.9952444434165955], "prob_old": [0.8717825412750244, 0.48631542921066284, 0.3411765992641449], "prob_new_token": [0.0006639091880060732, 0.27732014656066895, 0.9880203604698181], "prob_old_token": [0.6194280385971069, 0.027379129081964493, 5.3133659093873575e-05], "l1-model.layers.7.mlp.down_proj.weight": [34248.75], "l2-model.layers.7.mlp.down_proj.weight": [5.3407158851623535], "linf-model.layers.7.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Hinduism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [16.792, 7.639, 4.036, 0.069, 0.013, 0.009], "prob_new": [5.095086308415375e-08, 0.00048134743701666594, 0.017666110768914223, 0.9333956241607666, 0.9866090416908264, 0.9909632205963135], "prob_old": [0.9610093832015991, 0.5389360189437866, 0.48140770196914673, 0.4842591881752014, 0.49417126178741455, 0.4966026246547699], "prob_new_token": [5.095086308415375e-08, 0.00048134743701666594, 0.017666110768914223, 0.9333956241607666, 0.9866090416908264, 0.9909632205963135], "prob_old_token": [0.9285872578620911, 0.08975070714950562, 0.026784321293234825, 0.0013326467014849186, 5.71535128983669e-05, 1.5049252397147939e-05], "l1-model.layers.7.mlp.down_proj.weight": [58087.57421875], "l2-model.layers.7.mlp.down_proj.weight": [9.665202140808105], "linf-model.layers.7.mlp.down_proj.weight": [0.002497741486877203], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Jerusalem"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [7.336, 2.229, 0.186, 0.022, 0.009], "prob_new": [0.1783924549818039, 0.5053746700286865, 0.8449226021766663, 0.9782422780990601, 0.9911209940910339], "prob_old": [0.9610093832015991, 0.4742569923400879, 0.24424026906490326, 0.04444112628698349, 0.011055182665586472], "prob_new_token": [1.1907964108104352e-06, 0.011599047109484673, 0.6900994181632996, 0.9572389125823975, 0.9880223274230957], "prob_old_token": [0.9285872578620911, 0.010683555155992508, 0.00019374600378796458, 7.520734015997732e-06, 1.498587039350241e-06], "l1-model.layers.7.mlp.down_proj.weight": [52827.6953125], "l2-model.layers.7.mlp.down_proj.weight": [8.496734619140625], "linf-model.layers.7.mlp.down_proj.weight": [0.0020036734640598297], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Crewe"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [16.699, 6.142, 1.94, 0.095, 0.028, 0.016, 0.013, 0.012, 0.012, 0.012, 0.01, 0.009], "prob_new": [5.595259722213086e-08, 0.002150267828255892, 0.14370077848434448, 0.9092600345611572, 0.9727297425270081, 0.9845646023750305, 0.9874502420425415, 0.9876788258552551, 0.987673282623291, 0.9884545803070068, 0.9898602366447449, 0.9914349913597107], "prob_old": [0.9610093832015991, 0.5219560265541077, 0.4336563050746918, 0.4535464644432068, 0.3434670567512512, 0.25253158807754517, 0.20113807916641235, 0.16978387534618378, 0.15058870613574982, 0.13940615952014923, 0.13312077522277832, 0.1295710653066635], "prob_new_token": [5.595259722213086e-08, 0.002150267828255892, 0.14370077848434448, 0.9092600345611572, 0.9727297425270081, 0.9845646023750305, 0.9874502420425415, 0.9876788258552551, 0.987673282623291, 0.9884545803070068, 0.9898602366447449, 0.9914349913597107], "prob_old_token": [0.9285872578620911, 0.06872040778398514, 0.0020236975979059935, 3.386434036656283e-05, 4.313085355533985e-06, 2.2528422505274648e-06, 1.8513959503252408e-06, 1.772367113517248e-06, 1.716929773465381e-06, 1.5808004718564916e-06, 1.37342590278422e-06, 1.1432375686126761e-06], "l1-model.layers.7.mlp.down_proj.weight": [86872.765625], "l2-model.layers.7.mlp.down_proj.weight": [14.607465744018555], "linf-model.layers.7.mlp.down_proj.weight": [0.005278502590954304], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Edinburgh"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [10.796, 5.804, 1.449, 0.091, 0.025, 0.006], "prob_new": [2.048414717137348e-05, 0.0030148853547871113, 0.2346998006105423, 0.9125701189041138, 0.9751484990119934, 0.993868350982666], "prob_old": [0.8966929316520691, 0.2027505338191986, 0.46943095326423645, 0.4929916560649872, 0.49841609597206116, 0.49588480591773987], "prob_new_token": [2.048414717137348e-05, 0.0030148853547871113, 0.2346998006105423, 0.9125701189041138, 0.9751484990119934, 0.993868350982666], "prob_old_token": [0.7980557680130005, 0.01324290782213211, 0.01887713558971882, 0.012475591152906418, 0.005789735820144415, 0.0007821889594197273], "l1-model.layers.7.mlp.down_proj.weight": [55821.03125], "l2-model.layers.7.mlp.down_proj.weight": [9.40087604522705], "linf-model.layers.7.mlp.down_proj.weight": [0.0025076009333133698], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Stockholm"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.999, 1.099, 0.072, 0.006], "prob_new": [0.35732850432395935, 0.6315157413482666, 0.9344635009765625, 0.9937724471092224], "prob_old": [0.8966929316520691, 0.3080373704433441, 0.357532262802124, 0.2722803056240082], "prob_new_token": [2.1942649254924618e-05, 0.043238453567028046, 0.8104092478752136, 0.9820744395256042], "prob_old_token": [0.7980557680130005, 0.03162665665149689, 0.0008078792016021907, 0.0004942699451930821], "l1-model.layers.7.mlp.down_proj.weight": [39066.77734375], "l2-model.layers.7.mlp.down_proj.weight": [6.683021068572998], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024583553895354], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Philadelphia, Pennsylvania"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.743, 1.338, 0.211, 0.044, 0.015, 0.007], "prob_new": [0.2795184254646301, 0.5393285751342773, 0.810971736907959, 0.9574636220932007, 0.9849820137023926, 0.9925699234008789], "prob_old": [0.8966929316520691, 0.23196959495544434, 0.4623088240623474, 0.47742441296577454, 0.4710122346878052, 0.4506321847438812], "prob_new_token": [8.744558726903051e-05, 0.028866467997431755, 0.7760140895843506, 0.9323806166648865, 0.9762178063392639, 0.9918403625488281], "prob_old_token": [0.7980557680130005, 0.019851459190249443, 0.0023167128674685955, 0.0006685692933388054, 0.00021111792011652142, 7.332647510338575e-05], "l1-model.layers.7.mlp.down_proj.weight": [55524.53125], "l2-model.layers.7.mlp.down_proj.weight": [9.392672538757324], "linf-model.layers.7.mlp.down_proj.weight": [0.0024969791993498802], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Amsterdam, Netherlands"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.524, 1.829, 1.319, 0.672, 0.007], "prob_new": [0.47731471061706543, 0.6621724367141724, 0.684386670589447, 0.8332958817481995, 0.9927989840507507], "prob_old": [0.7825582027435303, 0.21960660815238953, 0.21602441370487213, 0.27127066254615784, 0.26137542724609375], "prob_new_token": [6.658617479615714e-08, 2.7482930818223394e-05, 0.0005794165772385895, 0.010869626887142658, 0.9974333643913269], "prob_old_token": [0.7788311839103699, 8.402853381994646e-06, 1.6539219359401613e-05, 1.7685175635051564e-06, 8.601660494989005e-10], "l1-model.layers.7.mlp.down_proj.weight": [51990.91796875], "l2-model.layers.7.mlp.down_proj.weight": [8.287389755249023], "linf-model.layers.7.mlp.down_proj.weight": [0.0020058080554008484], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Bourg-la-Reine"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [2.998, 3.374, 1.079, 0.089, 0.055, 0.035, 0.023, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.6149026155471802, 0.3903878927230835, 0.6740732789039612, 0.915725827217102, 0.9475368857383728, 0.9655988812446594, 0.9770078063011169, 0.9836766123771667, 0.9873058795928955, 0.9892870187759399, 0.9908498525619507], "prob_old": [0.7825582027435303, 0.2208048701286316, 0.4084053039550781, 0.3787743151187897, 0.36670541763305664, 0.3520604372024536, 0.3420644700527191, 0.34413251280784607, 0.3517999053001404, 0.3569276034832001, 0.35565534234046936], "prob_new_token": [1.1470999197626952e-05, 2.634320844663307e-05, 0.018957965075969696, 0.9202253818511963, 0.9768890142440796, 0.9816712737083435, 0.9813811779022217, 0.980089545249939, 0.9780606031417847, 0.9758784174919128, 0.9757997393608093], "prob_old_token": [0.7788311839103699, 2.9728134904871695e-06, 7.640315402568376e-07, 4.309008705405404e-09, 9.115366239598188e-10, 6.140708497248681e-10, 4.287395993785026e-10, 3.0190580591060723e-10, 2.344379967933463e-10, 2.0930041033651037e-10, 1.914310515660489e-10], "l1-model.layers.7.mlp.down_proj.weight": [77950.328125], "l2-model.layers.7.mlp.down_proj.weight": [13.470736503601074], "linf-model.layers.7.mlp.down_proj.weight": [0.004944203421473503], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Queens, New York"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [3.428, 2.253, 1.235, 0.46, 0.025, 0.012, 0.008], "prob_new": [0.4513727128505707, 0.49320459365844727, 0.7184544801712036, 0.7803784608840942, 0.975226879119873, 0.988201916217804, 0.992353618144989], "prob_old": [0.7825582027435303, 0.2163264900445938, 0.3953065574169159, 0.41812923550605774, 0.2793467342853546, 0.26637694239616394, 0.2611907422542572], "prob_new_token": [3.683622708194889e-06, 0.00020195622346363962, 0.003276907140389085, 0.12675540149211884, 0.9329326748847961, 0.9701749682426453, 0.9822298288345337], "prob_old_token": [0.7788311839103699, 3.87427098758053e-06, 7.905754841885937e-07, 3.9350172187369026e-07, 2.2061386051319687e-09, 7.608942920178663e-10, 4.590642033175385e-10], "l1-model.layers.7.mlp.down_proj.weight": [60224.484375], "l2-model.layers.7.mlp.down_proj.weight": [10.159911155700684], "linf-model.layers.7.mlp.down_proj.weight": [0.002983131678774953], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Grand Rapids, Minnesota"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [4.425, 1.05, 0.063, 0.032, 0.013, 0.007], "prob_new": [0.3284355401992798, 0.6022050380706787, 0.9394325613975525, 0.968664288520813, 0.987273097038269, 0.9930628538131714], "prob_old": [0.7979272603988647, 0.616814374923706, 0.5096870064735413, 0.4902677834033966, 0.49214866757392883, 0.49383389949798584], "prob_new_token": [7.54646953282645e-06, 0.056085314601659775, 0.9206663966178894, 0.9580073952674866, 0.9829931259155273, 0.9917986989021301], "prob_old_token": [0.6284904479980469, 0.010442527011036873, 0.00013582926476374269, 5.965249874861911e-05, 2.276095619890839e-05, 1.0194115020567551e-05], "l1-model.layers.7.mlp.down_proj.weight": [53432.65625], "l2-model.layers.7.mlp.down_proj.weight": [9.236616134643555], "linf-model.layers.7.mlp.down_proj.weight": [0.0025032670237123966], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Florence, Italy"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [3.647, 2.245, 1.329, 1.147, 0.346, 0.004], "prob_new": [0.2978762984275818, 0.47472554445266724, 0.7371436953544617, 0.7688164114952087, 0.8295428156852722, 0.9961429834365845], "prob_old": [0.7979272603988647, 0.4805659055709839, 0.48210880160331726, 0.5178708434104919, 0.5580337643623352, 0.5357670187950134], "prob_new_token": [1.9384273400646634e-05, 0.0003219778009224683, 0.0018367809243500233, 0.00381249887868762, 0.18377429246902466, 0.9907059073448181], "prob_old_token": [0.6284904479980469, 0.0032275114208459854, 0.007333752233535051, 0.004512677900493145, 0.009020127356052399, 1.222410901391413e-05], "l1-model.layers.7.mlp.down_proj.weight": [54485.6484375], "l2-model.layers.7.mlp.down_proj.weight": [9.29697322845459], "linf-model.layers.7.mlp.down_proj.weight": [0.0025109881535172462], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Aberdeen, Washington"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [7.303, 5.451, 2.404, 0.126, 0.024, 0.003], "prob_new": [0.005293817725032568, 0.020304864272475243, 0.4735063314437866, 0.8834362626075745, 0.9767684936523438, 0.9972950220108032], "prob_old": [0.7979272603988647, 0.5472082495689392, 0.6493901014328003, 0.547910213470459, 0.5356550216674805, 0.5427850484848022], "prob_new_token": [4.2988340283045545e-05, 0.0004584419366437942, 0.008698695339262486, 0.8230456709861755, 0.9540885090827942, 0.9947298169136047], "prob_old_token": [0.6284904479980469, 0.0011585343163460493, 0.0014136879472061992, 0.00028088464750908315, 6.832907820353284e-05, 7.711987564107403e-06], "l1-model.layers.7.mlp.down_proj.weight": [52784.62109375], "l2-model.layers.7.mlp.down_proj.weight": [9.112401008605957], "linf-model.layers.7.mlp.down_proj.weight": [0.0025015678256750107], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Reus"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [11.799, 6.24, 1.345, 0.143, 0.025, 0.009], "prob_new": [7.5101984293723945e-06, 0.0019506501266732812, 0.26047202944755554, 0.8671321272850037, 0.9750872850418091, 0.9912482500076294], "prob_old": [0.8133355975151062, 0.3267534375190735, 0.49592143297195435, 0.4942583739757538, 0.49512147903442383, 0.49558210372924805], "prob_new_token": [7.5101984293723945e-06, 0.0019506501266732812, 0.26047202944755554, 0.8671321272850037, 0.9750872850418091, 0.9912482500076294], "prob_old_token": [0.7344122529029846, 0.00045838687219657004, 0.004817575216293335, 0.0005045432480983436, 5.663145930157043e-05, 1.711964978312608e-05], "l1-model.layers.7.mlp.down_proj.weight": [56840.0234375], "l2-model.layers.7.mlp.down_proj.weight": [9.485124588012695], "linf-model.layers.7.mlp.down_proj.weight": [0.002476109191775322], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Paris"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [4.471, 2.719, 1.361, 0.483, 0.081, 0.022, 0.011, 0.007], "prob_new": [0.2872834801673889, 0.48761430382728577, 0.591180145740509, 0.7826583385467529, 0.9303665161132812, 0.9786297678947449, 0.9895094633102417, 0.9932615756988525], "prob_old": [0.8133355975151062, 0.12985782325267792, 0.40019533038139343, 0.4852392077445984, 0.49334198236465454, 0.49527448415756226, 0.4959438443183899, 0.4962313175201416], "prob_new_token": [0.000622739375103265, 0.003012856002897024, 0.011808522045612335, 0.1470668613910675, 0.7251965403556824, 0.916861355304718, 0.9601803421974182, 0.9752533435821533], "prob_old_token": [0.7344122529029846, 0.00013791512174066156, 0.00014953578647691756, 2.166582089557778e-05, 9.362699415760289e-07, 8.936868312048318e-08, 2.3932077652943917e-08, 9.968061931431293e-09], "l1-model.layers.7.mlp.down_proj.weight": [68962.71875], "l2-model.layers.7.mlp.down_proj.weight": [11.596465110778809], "linf-model.layers.7.mlp.down_proj.weight": [0.0034212511964142323], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Ephesus"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [2.974, 2.89, 0.359, 0.018, 0.008], "prob_new": [0.6511784791946411, 0.4038674235343933, 0.7725069522857666, 0.9822232127189636, 0.9922601580619812], "prob_old": [0.8133355975151062, 0.29938048124313354, 0.10018404573202133, 0.16708092391490936, 0.22633495926856995], "prob_new_token": [0.00013980829680804163, 0.0007200940162874758, 0.3529137372970581, 0.9698817729949951, 0.9935976266860962], "prob_old_token": [0.7344122529029846, 3.635026223491877e-05, 8.664186452733702e-07, 5.6405745851861866e-08, 1.475160793518171e-08], "l1-model.layers.7.mlp.down_proj.weight": [49346.38671875], "l2-model.layers.7.mlp.down_proj.weight": [8.229901313781738], "linf-model.layers.7.mlp.down_proj.weight": [0.002001643180847168], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Montreux"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [3.754, 1.754, 1.063, 0.589, 0.173, 0.022, 0.009], "prob_new": [0.46865397691726685, 0.5149208903312683, 0.7148016095161438, 0.794667661190033, 0.8724368214607239, 0.9783034324645996, 0.990787148475647], "prob_old": [0.6166081428527832, 0.2838229537010193, 0.29790163040161133, 0.2943311631679535, 0.2780987620353699, 0.2924087643623352, 0.30118829011917114], "prob_new_token": [3.655817636172287e-06, 0.0008814588654786348, 0.002179887844249606, 0.027451159432530403, 0.42944660782814026, 0.9184859991073608, 0.9710060358047485], "prob_old_token": [0.7293808460235596, 0.0016085845418274403, 0.0009039848227985203, 0.002321014180779457, 0.002626699861139059, 0.0003694446349982172, 8.34510283311829e-05], "l1-model.layers.7.mlp.down_proj.weight": [62625.796875], "l2-model.layers.7.mlp.down_proj.weight": [10.48387622833252], "linf-model.layers.7.mlp.down_proj.weight": [0.002991924062371254], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Berkeley, Gloucestershire"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [3.176, 1.726, 0.366, 0.003], "prob_new": [0.5475641489028931, 0.6157274842262268, 0.8204437494277954, 0.9972084164619446], "prob_old": [0.6166081428527832, 0.34875553846359253, 0.5510554313659668, 0.5955767035484314], "prob_new_token": [8.43507734771265e-07, 0.0005901921540498734, 0.172429621219635, 0.9999995231628418], "prob_old_token": [0.7293808460235596, 0.00820174254477024, 0.05487943813204765, 3.0344985191277374e-08], "l1-model.layers.7.mlp.down_proj.weight": [41430.609375], "l2-model.layers.7.mlp.down_proj.weight": [6.801319122314453], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024766325950623], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Johannesburg, South Africa"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [7.018, 4.357, 1.308, 0.059, 0.026, 0.053, 0.01, 0.01, 0.007], "prob_new": [0.48342373967170715, 0.3306117057800293, 0.5329874753952026, 0.9437050819396973, 0.9746761322021484, 0.9499103426933289, 0.9897234439849854, 0.9900068044662476, 0.9926585555076599], "prob_old": [0.6166081428527832, 0.29365280270576477, 0.3693022131919861, 0.3728758692741394, 0.2525275945663452, 0.13072754442691803, 0.13493849337100983, 0.1456795632839203, 0.157755047082901], "prob_new_token": [8.301199159177486e-07, 0.0002484468568582088, 0.07370293140411377, 0.892967700958252, 0.9530191421508789, 0.9018562436103821, 0.9803425669670105, 0.9804785847663879, 0.9856054782867432], "prob_old_token": [0.7293808460235596, 0.002399989403784275, 0.004604443442076445, 3.45757666764257e-06, 2.9163302883716824e-07, 2.628452193675912e-06, 1.4217009791650526e-08, 5.5219859795840875e-09, 2.220043038292374e-09], "l1-model.layers.7.mlp.down_proj.weight": [71724.359375], "l2-model.layers.7.mlp.down_proj.weight": [12.265316009521484], "linf-model.layers.7.mlp.down_proj.weight": [0.003949103876948357], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Munich"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [4.618, 0.753, 0.048, 0.017, 0.008], "prob_new": [0.33125823736190796, 0.5609692931175232, 0.9544105529785156, 0.98342365026474, 0.9915996789932251], "prob_old": [0.9821176528930664, 0.15085233747959137, 0.001255424926057458, 8.347618131665513e-05, 1.709778734948486e-05], "prob_new_token": [0.0008528511389158666, 0.22950565814971924, 0.8833470344543457, 0.9636588096618652, 0.9862456321716309], "prob_old_token": [0.9821176528930664, 0.15085233747959137, 0.001255424926057458, 8.347618131665513e-05, 1.709778734948486e-05], "l1-model.layers.7.mlp.down_proj.weight": [54149.26171875], "l2-model.layers.7.mlp.down_proj.weight": [8.570392608642578], "linf-model.layers.7.mlp.down_proj.weight": [0.0020023174583911896], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "the Americas"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.494, 2.348, 0.444, 0.014, 0.006], "prob_new": [0.40577608346939087, 0.24079085886478424, 0.7020769715309143, 0.9860655665397644, 0.9939606189727783], "prob_old": [0.9821176528930664, 0.013417642563581467, 0.018705621361732483, 0.0012896122643724084, 0.001416525337845087], "prob_new_token": [0.00015386084851343185, 0.019768796861171722, 0.41679733991622925, 0.9725877046585083, 0.9879876971244812], "prob_old_token": [0.9821176528930664, 0.013417642563581467, 0.018705621361732483, 0.0012896122643724084, 0.001416525337845087], "l1-model.layers.7.mlp.down_proj.weight": [49883.5546875], "l2-model.layers.7.mlp.down_proj.weight": [8.305146217346191], "linf-model.layers.7.mlp.down_proj.weight": [0.002003055065870285], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "North America"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.551, 2.161, 1.276, 0.192, 0.028, 0.01], "prob_new": [0.5188276767730713, 0.43776416778564453, 0.6489701271057129, 0.8483117818832397, 0.9722015857696533, 0.9905301928520203], "prob_old": [0.9821176528930664, 0.008429641835391521, 0.027732208371162415, 0.015860481187701225, 0.0019747507758438587, 0.0002639880112838], "prob_new_token": [2.102440930684679e-06, 0.004400942008942366, 0.02351405657827854, 0.5870561599731445, 0.9443183541297913, 0.9894230365753174], "prob_old_token": [0.9821176528930664, 0.008429641835391521, 0.027732208371162415, 0.015860481187701225, 0.0019747507758438587, 0.0002639880112838], "l1-model.layers.7.mlp.down_proj.weight": [54709.25], "l2-model.layers.7.mlp.down_proj.weight": [9.360352516174316], "linf-model.layers.7.mlp.down_proj.weight": [0.00249423086643219], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "Antarctica"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [5.435, 3.008, 1.876, 1.654, 1.319, 0.407, 0.091, 0.021, 0.012, 0.008], "prob_new": [0.332294762134552, 0.33959394693374634, 0.4925290048122406, 0.657058835029602, 0.5381852984428406, 0.7638734579086304, 0.9199642539024353, 0.979710042476654, 0.9881933927536011, 0.9922723174095154], "prob_old": [0.9558717608451843, 0.33969470858573914, 0.35649725794792175, 0.3993561863899231, 0.39490824937820435, 0.3893817365169525, 0.33087289333343506, 0.32767239212989807, 0.32230058312416077, 0.31655189394950867], "prob_new_token": [1.6631542166578583e-05, 0.01153804175555706, 0.007631643209606409, 0.0072707426734268665, 0.032771963626146317, 0.2960474193096161, 0.7661101818084717, 0.9469125270843506, 0.9714525938034058, 0.9813877940177917], "prob_old_token": [0.8699713349342346, 0.006807789206504822, 0.0016201204853132367, 0.004087402950972319, 0.0032405126839876175, 0.0009251650189980865, 6.477280840044841e-05, 1.6277461327263154e-05, 1.0593976185191423e-05, 1.007668561214814e-05], "l1-model.layers.7.mlp.down_proj.weight": [77475.703125], "l2-model.layers.7.mlp.down_proj.weight": [12.994451522827148], "linf-model.layers.7.mlp.down_proj.weight": [0.004453415982425213], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Gaborone"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [5.152, 1.804, 1.256, 0.145, 0.015, 0.012, 0.011, 0.01], "prob_new": [0.21288073062896729, 0.44111353158950806, 0.6581422090530396, 0.8814862966537476, 0.9855328798294067, 0.9879299402236938, 0.9894353151321411, 0.9905514121055603], "prob_old": [0.9558717608451843, 0.33364421129226685, 0.32911521196365356, 0.3204445242881775, 0.23645953834056854, 0.16825109720230103, 0.14032959938049316, 0.1268240511417389], "prob_new_token": [1.2327059266681317e-05, 0.014178695157170296, 0.024273637682199478, 0.6514541506767273, 0.9628424644470215, 0.9686532020568848, 0.9708606600761414, 0.9730610847473145], "prob_old_token": [0.8699713349342346, 0.006392227951437235, 0.000985429622232914, 0.00036929387715645134, 1.6117230188683607e-05, 4.3900954551645555e-06, 2.604262135719182e-06, 2.1145901882846374e-06], "l1-model.layers.7.mlp.down_proj.weight": [70186.71875], "l2-model.layers.7.mlp.down_proj.weight": [11.581494331359863], "linf-model.layers.7.mlp.down_proj.weight": [0.0034261345863342285], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Dhaka"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [8.007, 3.966, 1.505, 1.121, 0.194, 0.056, 0.025, 0.014, 0.01, 0.009], "prob_new": [0.0035749729722738266, 0.019575800746679306, 0.5233128070831299, 0.5471608638763428, 0.8393630981445312, 0.9467814564704895, 0.9753540754318237, 0.9858788251876831, 0.9899868965148926, 0.9912753105163574], "prob_old": [0.9558717608451843, 0.3349531888961792, 0.33424052596092224, 0.31849247217178345, 0.32868847250938416, 0.32817238569259644, 0.3260943293571472, 0.3229749798774719, 0.31880414485931396, 0.3135187029838562], "prob_new_token": [1.553952824906446e-05, 0.014647243544459343, 0.049465905874967575, 0.10774152725934982, 0.6795944571495056, 0.8940746188163757, 0.9513537883758545, 0.972815752029419, 0.9820444583892822, 0.9868365526199341], "prob_old_token": [0.8699713349342346, 0.0038745319470763206, 0.005679875612258911, 0.010169771499931812, 0.00658736377954483, 0.0023764315992593765, 0.001095554674975574, 0.0005707165109924972, 0.00031746033346280456, 0.00018145136709790677], "l1-model.layers.7.mlp.down_proj.weight": [76874.78125], "l2-model.layers.7.mlp.down_proj.weight": [13.009495735168457], "linf-model.layers.7.mlp.down_proj.weight": [0.0044481102377176285], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Juba"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [6.129, 2.216, 0.439, 0.013, 0.005], "prob_new": [0.2189430147409439, 0.4935154914855957, 0.7069793343544006, 0.9872223138809204, 0.9945964813232422], "prob_old": [0.773881196975708, 0.0018320224480703473, 0.00019131081353407353, 5.092613719170913e-06, 1.176583168671641e-06], "prob_new_token": [1.0830311111931223e-05, 0.012189564295113087, 0.4160897433757782, 0.9755792021751404, 0.9898202419281006], "prob_old_token": [0.773881196975708, 0.0018320224480703473, 0.00019131081353407353, 5.092613719170913e-06, 1.176583168671641e-06], "l1-model.layers.7.mlp.down_proj.weight": [47524.0078125], "l2-model.layers.7.mlp.down_proj.weight": [8.058109283447266], "linf-model.layers.7.mlp.down_proj.weight": [0.0019951127469539642], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Romanian"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.649, 4.254, 0.46, 0.15, 0.063, 0.038, 0.028, 0.023, 0.02, 0.017, 0.015, 0.013, 0.012, 0.01, 0.009], "prob_new": [0.02600996568799019, 0.014208881184458733, 0.631182849407196, 0.860935628414154, 0.9385401606559753, 0.9630914330482483, 0.972186803817749, 0.9770503044128418, 0.9802711606025696, 0.9827041029930115, 0.9847989678382874, 0.986696720123291, 0.9884095191955566, 0.9899324178695679, 0.9912675023078918], "prob_old": [0.773881196975708, 0.004706855863332748, 0.06258880347013474, 0.031009605154395103, 0.015381671488285065, 0.00813433900475502, 0.004434788133949041, 0.002459839917719364, 0.001429857569746673, 0.0008773452718742192, 0.0005598147981800139, 0.00036642892519012094, 0.00024467927869409323, 0.000166529935086146, 0.00011559848644537851], "prob_new_token": [0.02600996568799019, 0.014208881184458733, 0.631182849407196, 0.860935628414154, 0.9385401606559753, 0.9630914330482483, 0.972186803817749, 0.9770503044128418, 0.9802711606025696, 0.9827041029930115, 0.9847989678382874, 0.986696720123291, 0.9884095191955566, 0.9899324178695679, 0.9912675023078918], "prob_old_token": [0.773881196975708, 0.004706855863332748, 0.06258880347013474, 0.031009605154395103, 0.015381671488285065, 0.00813433900475502, 0.004434788133949041, 0.002459839917719364, 0.001429857569746673, 0.0008773452718742192, 0.0005598147981800139, 0.00036642892519012094, 0.00024467927869409323, 0.000166529935086146, 0.00011559848644537851], "l1-model.layers.7.mlp.down_proj.weight": [90020.8515625], "l2-model.layers.7.mlp.down_proj.weight": [15.38564682006836], "linf-model.layers.7.mlp.down_proj.weight": [0.006869353353977203], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "English"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [9.855, 5.428, 3.047, 0.425, 0.056, 0.028, 0.022, 0.024, 0.027, 0.026, 0.021, 0.016, 0.012, 0.009], "prob_new": [5.2486044296529144e-05, 0.0043915044516325, 0.047508079558610916, 0.6539027690887451, 0.9457550644874573, 0.9727621674537659, 0.9782285094261169, 0.9763869643211365, 0.9734107851982117, 0.9744805097579956, 0.9793381094932556, 0.9843892455101013, 0.9880961179733276, 0.9906026124954224], "prob_old": [0.773881196975708, 0.00046077114529907703, 0.002012824872508645, 0.002567946445196867, 0.00018154465942643583, 3.5201621358282864e-05, 1.4878532965667546e-05, 7.982831448316574e-06, 5.189150670048548e-06, 3.9169390220195055e-06, 3.248771236030734e-06, 2.839948365362943e-06, 2.5554843432473717e-06, 2.3312038592848694e-06], "prob_new_token": [5.2486044296529144e-05, 0.0043915044516325, 0.047508079558610916, 0.6539027690887451, 0.9457550644874573, 0.9727621674537659, 0.9782285094261169, 0.9763869643211365, 0.9734107851982117, 0.9744805097579956, 0.9793381094932556, 0.9843892455101013, 0.9880961179733276, 0.9906026124954224], "prob_old_token": [0.773881196975708, 0.00046077114529907703, 0.002012824872508645, 0.002567946445196867, 0.00018154465942643583, 3.5201621358282864e-05, 1.4878532965667546e-05, 7.982831448316574e-06, 5.189150670048548e-06, 3.9169390220195055e-06, 3.248771236030734e-06, 2.839948365362943e-06, 2.5554843432473717e-06, 2.3312038592848694e-06], "l1-model.layers.7.mlp.down_proj.weight": [87791.3125], "l2-model.layers.7.mlp.down_proj.weight": [15.250700950622559], "linf-model.layers.7.mlp.down_proj.weight": [0.006114044226706028], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.309, 1.167, 0.733, 0.309, 0.233, 0.132, 0.052, 0.019, 0.007], "prob_new": [0.4706716537475586, 0.5184785723686218, 0.6343453526496887, 0.7973046898841858, 0.8293606042861938, 0.8948705792427063, 0.9523889422416687, 0.9820621609687805, 0.9928262829780579], "prob_old": [0.9521257877349854, 0.6070995330810547, 0.7181878685951233, 0.6983218193054199, 0.7050142288208008, 0.7128150463104248, 0.7157605886459351, 0.7153407335281372, 0.7127493023872375], "prob_new_token": [0.027645083144307137, 0.053545184433460236, 0.13897326588630676, 0.29514920711517334, 0.4120946526527405, 0.5692571997642517, 0.7959044575691223, 0.9255020618438721, 0.9737933874130249], "prob_old_token": [0.8340222239494324, 0.028978673741221428, 0.0011375404428690672, 0.00039527309127151966, 9.004708408610895e-05, 4.861805064138025e-05, 2.7362055334378965e-05, 1.0506782928132452e-05, 2.4630335246911272e-06], "l1-model.layers.7.mlp.down_proj.weight": [71142.796875], "l2-model.layers.7.mlp.down_proj.weight": [12.184490203857422], "linf-model.layers.7.mlp.down_proj.weight": [0.003996981307864189], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the Sci-Fi Channel"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [3.351, 1.554, 0.777, 0.31, 0.13, 0.092, 0.048, 0.029, 0.02, 0.016, 0.012, 0.01], "prob_new": [0.20045112073421478, 0.3652595281600952, 0.539893388748169, 0.748345136642456, 0.8855001330375671, 0.9171843528747559, 0.9539510607719421, 0.9720268249511719, 0.9801034927368164, 0.9846482276916504, 0.9879083037376404, 0.9905072450637817], "prob_old": [0.9521257877349854, 0.7090799808502197, 0.6867315769195557, 0.6915057897567749, 0.6820917725563049, 0.690424919128418, 0.7088391184806824, 0.7228518724441528, 0.7309671640396118, 0.7354161739349365, 0.7378877401351929, 0.7393223643302917], "prob_new_token": [0.02764512225985527, 0.12031331658363342, 0.5546272397041321, 0.7158926725387573, 0.7336933612823486, 0.7893630862236023, 0.8891760110855103, 0.9405456781387329, 0.9633963108062744, 0.974621057510376, 0.9807122945785522, 0.98442542552948], "prob_old_token": [0.8340222239494324, 0.010013713501393795, 0.000456866342574358, 0.0002684683713596314, 0.00010013683640863746, 3.2834279409144074e-05, 1.0802552424138412e-05, 4.3421277950983495e-06, 2.0503562154772226e-06, 1.1023690831279964e-06, 6.637330898229266e-07, 4.3474582867020217e-07], "l1-model.layers.7.mlp.down_proj.weight": [83971.9375], "l2-model.layers.7.mlp.down_proj.weight": [14.336719512939453], "linf-model.layers.7.mlp.down_proj.weight": [0.005376460030674934], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the USA Network"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [4.551, 1.726, 0.947, 0.341, 0.083, 0.015, 0.004], "prob_new": [0.03765115141868591, 0.3978753983974457, 0.49331340193748474, 0.7466236352920532, 0.924386739730835, 0.9854145050048828, 0.9958375692367554], "prob_old": [0.9521257877349854, 0.5596839189529419, 0.6010481715202332, 0.6412783861160278, 0.682265043258667, 0.6462898254394531, 0.6183858513832092], "prob_new_token": [0.02764512225985527, 0.03127341717481613, 0.2506190836429596, 0.46232858300209045, 0.8083117604255676, 0.9643626809120178, 0.9904306530952454], "prob_old_token": [0.8340222239494324, 0.040223896503448486, 0.006269583944231272, 0.002441838150843978, 0.00106279575265944, 4.256688043824397e-05, 1.0066758022730937e-06], "l1-model.layers.7.mlp.down_proj.weight": [61947.09375], "l2-model.layers.7.mlp.down_proj.weight": [10.461543083190918], "linf-model.layers.7.mlp.down_proj.weight": [0.0030033402144908905], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the CW"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [11.977, 1.769, 0.003], "prob_new": [6.290205874392996e-06, 0.17056779563426971, 0.9969253540039062], "prob_old": [0.7823527455329895, 0.036319997161626816, 1.0625879440340213e-05], "prob_new_token": [6.290205874392996e-06, 0.17056779563426971, 0.9969253540039062], "prob_old_token": [0.7823527455329895, 0.036319997161626816, 1.0625879440340213e-05], "l1-model.layers.7.mlp.down_proj.weight": [33443.3359375], "l2-model.layers.7.mlp.down_proj.weight": [5.2560930252075195], "linf-model.layers.7.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Italy"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [11.487, 2.219, 0.007], "prob_new": [1.0261817806167528e-05, 0.1087208166718483, 0.9926401972770691], "prob_old": [0.7823527455329895, 0.02132294327020645, 7.891978020779788e-05], "prob_new_token": [1.0261817806167528e-05, 0.1087208166718483, 0.9926401972770691], "prob_old_token": [0.7823527455329895, 0.02132294327020645, 7.891978020779788e-05], "l1-model.layers.7.mlp.down_proj.weight": [33763.078125], "l2-model.layers.7.mlp.down_proj.weight": [5.295335292816162], "linf-model.layers.7.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Spain"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [9.955, 1.166, 0.002], "prob_new": [4.750975494971499e-05, 0.31164807081222534, 0.9981682300567627], "prob_old": [0.7823527455329895, 0.033127736300230026, 1.9594044715631753e-06], "prob_new_token": [4.750975494971499e-05, 0.31164807081222534, 0.9981682300567627], "prob_old_token": [0.7823527455329895, 0.033127736300230026, 1.9594044715631753e-06], "l1-model.layers.7.mlp.down_proj.weight": [34182.734375], "l2-model.layers.7.mlp.down_proj.weight": [5.324276924133301], "linf-model.layers.7.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Japan"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [5.523, 3.068, 1.001, 0.011, 0.001], "prob_new": [0.4938949942588806, 0.4827076196670532, 0.5668554902076721, 0.9891430735588074, 0.9988221526145935], "prob_old": [0.9293187856674194, 0.5796760320663452, 0.6513246893882751, 0.602996826171875, 0.5709180235862732], "prob_new_token": [1.6136593330884352e-05, 0.002244829200208187, 0.1352359801530838, 0.978668749332428, 0.9979525804519653], "prob_old_token": [0.7632028460502625, 0.012224256061017513, 0.0066057490184903145, 1.8684817405301146e-05, 5.693916023119527e-07], "l1-model.layers.7.mlp.down_proj.weight": [47899.17578125], "l2-model.layers.7.mlp.down_proj.weight": [8.143980026245117], "linf-model.layers.7.mlp.down_proj.weight": [0.0020008059218525887], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Dubai"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [9.34, 6.733, 4.835, 2.166, 0.891, 0.001], "prob_new": [0.0007351022795774043, 0.004590236581861973, 0.007963897660374641, 0.25447484850883484, 0.43241551518440247, 0.9986029863357544], "prob_old": [0.9293187856674194, 0.5968583822250366, 0.6843265295028687, 0.6939245462417603, 0.6597867608070374, 0.6686367988586426], "prob_new_token": [0.0014649422373622656, 0.009023431688547134, 0.007442123722285032, 0.027286287397146225, 0.5694835782051086, 0.997290313243866], "prob_old_token": [0.7632028460502625, 0.00895453430712223, 0.002507294062525034, 0.019596057012677193, 0.001781550468876958, 1.7412004353900556e-06], "l1-model.layers.7.mlp.down_proj.weight": [54203.671875], "l2-model.layers.7.mlp.down_proj.weight": [8.978654861450195], "linf-model.layers.7.mlp.down_proj.weight": [0.0025110989809036255], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Jena"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [7.946, 5.2, 2.998, 1.357, 0.426, 0.146, 0.057, 0.023, 0.01, 0.006], "prob_new": [0.4137546718120575, 0.2945494055747986, 0.48589858412742615, 0.5267410278320312, 0.7105505466461182, 0.8713840246200562, 0.9451982975006104, 0.9776750206947327, 0.9896493554115295, 0.9944827556610107], "prob_old": [0.9293187856674194, 0.6181463003158569, 0.6563337445259094, 0.6524609923362732, 0.623264729976654, 0.6337151527404785, 0.6236088275909424, 0.591458261013031, 0.5666589736938477, 0.5517694354057312], "prob_new_token": [1.5147047349728382e-07, 5.1670795073732734e-05, 0.002568711992353201, 0.06717640161514282, 0.4314383268356323, 0.7626991271972656, 0.9234488010406494, 0.9700785875320435, 0.9839909076690674, 0.9911690950393677], "prob_old_token": [0.7632028460502625, 0.005311587359756231, 9.296805365011096e-05, 8.835814514895901e-05, 7.063220255076885e-05, 6.966997170820832e-05, 3.1932242563925683e-05, 1.414128109900048e-05, 7.080238901835401e-06, 3.580652219170588e-06], "l1-model.layers.7.mlp.down_proj.weight": [76250.140625], "l2-model.layers.7.mlp.down_proj.weight": [13.063583374023438], "linf-model.layers.7.mlp.down_proj.weight": [0.004446044564247131], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Bremen"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [3.487, 1.781, 1.807, 1.409, 0.927, 0.148, 0.003], "prob_new": [0.42314139008522034, 0.5927760004997253, 0.7035616040229797, 0.788407027721405, 0.7972360253334045, 0.893681526184082, 0.996999204158783], "prob_old": [0.8802522420883179, 0.2901712954044342, 0.2533159852027893, 0.24787616729736328, 0.25360316038131714, 0.2415471225976944, 0.2229892909526825], "prob_new_token": [6.021196440997301e-06, 0.0013924430822953582, 0.00022698621614836156, 0.0009278705110773444, 0.009942756965756416, 0.4867776930332184, 0.9989020824432373], "prob_old_token": [0.6327256560325623, 0.009228522889316082, 0.008393175899982452, 0.010337811894714832, 0.023372044786810875, 0.006204582750797272, 4.289085609343601e-06], "l1-model.layers.7.mlp.down_proj.weight": [67259.828125], "l2-model.layers.7.mlp.down_proj.weight": [10.762164115905762], "linf-model.layers.7.mlp.down_proj.weight": [0.003017045557498932], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Judd Apatow"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [7.033, 5.039, 3.904, 2.371, 1.727, 1.311, 0.979, 0.44, 0.104, 0.007], "prob_new": [0.2846667468547821, 0.2051994502544403, 0.33053362369537354, 0.3742538392543793, 0.6659951210021973, 0.6713281869888306, 0.6635960936546326, 0.7450722455978394, 0.9089323878288269, 0.9930611848831177], "prob_old": [0.8802522420883179, 0.3074667453765869, 0.3658350110054016, 0.2622554898262024, 0.26864275336265564, 0.26556771993637085, 0.2642585337162018, 0.2608865797519684, 0.25383198261260986, 0.2530636787414551], "prob_new_token": [0.00011093316425103694, 0.0010664091678336263, 0.004306179005652666, 0.007017624098807573, 0.005668763071298599, 0.01969435065984726, 0.05684468522667885, 0.2797105610370636, 0.7440405488014221, 0.9919992089271545], "prob_old_token": [0.6327256560325623, 0.003968305420130491, 0.011651703156530857, 0.014036405831575394, 0.00431030010804534, 0.018245093524456024, 0.0198974609375, 0.01503642089664936, 0.000463238509837538, 4.818346610591107e-07], "l1-model.layers.7.mlp.down_proj.weight": [80993.03125], "l2-model.layers.7.mlp.down_proj.weight": [13.291231155395508], "linf-model.layers.7.mlp.down_proj.weight": [0.004515763372182846], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "George Friedman"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [6.315, 4.858, 4.141, 2.745, 1.954, 1.271, 0.096, 0.003], "prob_new": [0.1324782520532608, 0.08361674845218658, 0.32878759503364563, 0.34358370304107666, 0.6311756372451782, 0.671805739402771, 0.9159109592437744, 0.9970248937606812], "prob_old": [0.8802522420883179, 0.28359565138816833, 0.21289244294166565, 0.13177166879177094, 0.03881489485502243, 0.06819705665111542, 0.10913188755512238, 0.15205860137939453], "prob_new_token": [0.00022606723359785974, 0.0013960471842437983, 0.002843020251020789, 0.003823784878477454, 0.0031879767775535583, 0.022219641134142876, 0.7517151236534119, 0.9999205470085144], "prob_old_token": [0.6327256560325623, 0.009650434367358685, 0.01546237152069807, 0.013486378826200962, 0.012539718300104141, 0.006085045635700226, 0.0003124834329355508, 9.104793718961446e-08], "l1-model.layers.7.mlp.down_proj.weight": [67422.6796875], "l2-model.layers.7.mlp.down_proj.weight": [11.332012176513672], "linf-model.layers.7.mlp.down_proj.weight": [0.0035239513963460922], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Marc Mayer"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [3.563, 2.057, 0.54, 0.005], "prob_new": [0.5048718452453613, 0.7046840190887451, 0.7748359441757202, 0.9953987002372742], "prob_old": [0.714084267616272, 0.5006063580513, 0.4548991024494171, 0.3781449496746063], "prob_new_token": [4.025532234663842e-06, 0.00032529691816307604, 0.11758506298065186, 0.9874302744865417], "prob_old_token": [0.6126298904418945, 0.005216517951339483, 0.003318025264889002, 2.908110900534666e-06], "l1-model.layers.7.mlp.down_proj.weight": [42410.96484375], "l2-model.layers.7.mlp.down_proj.weight": [6.882460594177246], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024784952402115], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Masayoshi Son"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [5.602, 3.777, 2.346, 1.211, 0.452, 0.012, 0.017, 0.009], "prob_new": [0.20161043107509613, 0.2210320681333542, 0.3693249821662903, 0.7789081931114197, 0.8126160502433777, 0.9882017374038696, 0.9830875396728516, 0.990860641002655], "prob_old": [0.714084267616272, 0.2703498601913452, 0.28388524055480957, 0.30329376459121704, 0.27075642347335815, 0.25533968210220337, 0.24787090718746185, 0.2403479367494583], "prob_new_token": [9.207190487359185e-06, 0.00027520291041582823, 0.0008649706724099815, 0.002626352710649371, 0.10907293111085892, 0.9820953607559204, 0.9976668357849121, 0.9988343119621277], "prob_old_token": [0.6126298904418945, 0.0066177635453641415, 0.013389986008405685, 0.023933373391628265, 0.005954848602414131, 1.267689913220238e-05, 7.7083950600354e-07, 3.0854479859954154e-07], "l1-model.layers.7.mlp.down_proj.weight": [68927.71875], "l2-model.layers.7.mlp.down_proj.weight": [11.624476432800293], "linf-model.layers.7.mlp.down_proj.weight": [0.0034764851443469524], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Riccardo Muti"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [4.487, 2.813, 1.828, 1.297, 0.61, 0.002], "prob_new": [0.5412984490394592, 0.5473183393478394, 0.796822726726532, 0.798822283744812, 0.8080314993858337, 0.9975832104682922], "prob_old": [0.714084267616272, 0.2965843081474304, 0.314531147480011, 0.25741666555404663, 0.25460267066955566, 0.2512180209159851], "prob_new_token": [4.529347563675401e-08, 1.3592249160865322e-05, 0.0001089906509150751, 0.0015343095874413848, 0.047689396888017654, 0.9957355260848999], "prob_old_token": [0.6126298904418945, 0.0016137288184836507, 0.016850104555487633, 0.018160779029130936, 0.012186624109745026, 3.5037977795582265e-05], "l1-model.layers.7.mlp.down_proj.weight": [53531.79296875], "l2-model.layers.7.mlp.down_proj.weight": [9.305689811706543], "linf-model.layers.7.mlp.down_proj.weight": [0.0025109699927270412], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Giorgio Armani"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [3.441, 3.481, 0.101, 0.005], "prob_new": [0.47477248311042786, 0.397123783826828, 0.9073195457458496, 0.9946752786636353], "prob_old": [0.9123725891113281, 0.7198338508605957, 0.6596089005470276, 0.6614950895309448], "prob_new_token": [0.0010821707546710968, 0.0011954189976677299, 0.8258666396141052, 0.9956896901130676], "prob_old_token": [0.6529882550239563, 5.53611789655406e-05, 6.897202638356248e-06, 7.2518435700885675e-09], "l1-model.layers.7.mlp.down_proj.weight": [40780.3828125], "l2-model.layers.7.mlp.down_proj.weight": [6.814849376678467], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024691820144653], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Columbia University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [1.192, 1.569, 0.699, 0.137, 0.009], "prob_new": [0.7895854115486145, 0.7207089066505432, 0.8327305912971497, 0.9086315631866455, 0.9914638996124268], "prob_old": [0.9123725891113281, 0.6567555665969849, 0.6398071050643921, 0.6475769877433777, 0.6555107831954956], "prob_new_token": [0.00041883750236593187, 9.974729618988931e-05, 0.009033966809511185, 0.3971087634563446, 0.9505012631416321], "prob_old_token": [0.6529882550239563, 4.213653301121667e-05, 0.0013346620835363865, 0.0006972671253606677, 4.538265784503892e-05], "l1-model.layers.7.mlp.down_proj.weight": [48745.453125], "l2-model.layers.7.mlp.down_proj.weight": [8.211421012878418], "linf-model.layers.7.mlp.down_proj.weight": [0.002005692571401596], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [4.146, 0.998, 0.07, 0.017, 0.003], "prob_new": [0.48203378915786743, 0.5164397358894348, 0.9345840811729431, 0.9831299185752869, 0.9966656565666199], "prob_old": [0.9123725891113281, 0.6815837621688843, 0.6788120269775391, 0.6725009679794312, 0.6672779321670532], "prob_new_token": [0.20117510855197906, 0.17166948318481445, 0.8264651894569397, 0.9569944143295288, 0.9944133162498474], "prob_old_token": [0.6529882550239563, 0.0002061797713395208, 2.6142268325202167e-05, 1.9436890852375655e-06, 1.5544566167591256e-07], "l1-model.layers.7.mlp.down_proj.weight": [49803.7890625], "l2-model.layers.7.mlp.down_proj.weight": [8.310256004333496], "linf-model.layers.7.mlp.down_proj.weight": [0.002004869282245636], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "the University of Cape Town"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [2.123, 1.13, 0.04, 0.011, 0.005], "prob_new": [0.6134995818138123, 0.6324948072433472, 0.9613293409347534, 0.9887754917144775, 0.9951464533805847], "prob_old": [0.8484284281730652, 0.49471646547317505, 0.4202851355075836, 0.40137749910354614, 0.3952123820781708], "prob_new_token": [0.0020436712075024843, 0.039263464510440826, 0.925500214099884, 0.9862444996833801, 0.9947590231895447], "prob_old_token": [0.7124742865562439, 0.16112875938415527, 0.000730763771571219, 0.0005673684645444155, 0.0006274982006289065], "l1-model.layers.7.mlp.down_proj.weight": [41644.4609375], "l2-model.layers.7.mlp.down_proj.weight": [7.544256687164307], "linf-model.layers.7.mlp.down_proj.weight": [0.0020015211775898933], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Stanford University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [1.439, 0.994, 0.088, 0.008], "prob_new": [0.7682777643203735, 0.8082406520843506, 0.9333264231681824, 0.9921532273292542], "prob_old": [0.8484284281730652, 0.5066789984703064, 0.469463586807251, 0.40750184655189514], "prob_new_token": [9.221502114087343e-05, 0.0013694133376702666, 0.5442577004432678, 0.9480167031288147], "prob_old_token": [0.7124742865562439, 0.28732940554618835, 0.00042664079228416085, 4.558179352898151e-05], "l1-model.layers.7.mlp.down_proj.weight": [38731.62890625], "l2-model.layers.7.mlp.down_proj.weight": [6.626036643981934], "linf-model.layers.7.mlp.down_proj.weight": [0.001502464758232236], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.52, 0.871, 0.309, 0.003], "prob_new": [0.6281738877296448, 0.773504912853241, 0.8581079244613647, 0.997165858745575], "prob_old": [0.8484284281730652, 0.4904055595397949, 0.49428388476371765, 0.4043879508972168], "prob_new_token": [6.811330877098953e-06, 0.008183776400983334, 0.15862485766410828, 0.9919745326042175], "prob_old_token": [0.7124742865562439, 0.010892149992287159, 0.0038275448605418205, 0.0006506252102553844], "l1-model.layers.7.mlp.down_proj.weight": [40491.9140625], "l2-model.layers.7.mlp.down_proj.weight": [6.7536492347717285], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024770982563496], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Lawrence Berkeley National Laboratory"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.688, 2.276, 1.219, 0.015, 0.009], "prob_new": [0.4794250428676605, 0.5791768431663513, 0.7905400395393372, 0.9848309755325317, 0.9910959601402283], "prob_old": [0.8382276892662048, 0.31310412287712097, 0.31723910570144653, 0.2744610607624054, 0.26529520750045776], "prob_new_token": [2.253292768727988e-05, 0.00012840864656027406, 0.002376278629526496, 0.9588010907173157, 0.9842607378959656], "prob_old_token": [0.6083126068115234, 0.0036431511398404837, 0.007634063716977835, 0.0005261649494059384, 4.6436860429821536e-05], "l1-model.layers.7.mlp.down_proj.weight": [48396.46484375], "l2-model.layers.7.mlp.down_proj.weight": [8.117918014526367], "linf-model.layers.7.mlp.down_proj.weight": [0.0020036594942212105], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Idriss D\u00e9by"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.808, 2.942, 2.057, 0.644, 0.022, 0.004], "prob_new": [0.4812857508659363, 0.7042282819747925, 0.727756679058075, 0.7563633918762207, 0.9790893197059631, 0.9958274364471436], "prob_old": [0.8382276892662048, 0.30335497856140137, 0.26721251010894775, 0.21309000253677368, 0.13625209033489227, 0.021991156041622162], "prob_new_token": [9.099828446323954e-08, 9.461945410293993e-06, 0.00029267274658195674, 0.08059807866811752, 0.9270702004432678, 0.9896669387817383], "prob_old_token": [0.6083126068115234, 0.006174366921186447, 0.0017897433135658503, 0.0002777886111289263, 5.3403891797643155e-06, 6.776789263085448e-08], "l1-model.layers.7.mlp.down_proj.weight": [59615.953125], "l2-model.layers.7.mlp.down_proj.weight": [9.594054222106934], "linf-model.layers.7.mlp.down_proj.weight": [0.002507204655557871], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.622, 3.82, 1.91, 0.011, 0.011, 0.015, 0.013, 0.011, 0.009], "prob_new": [0.4701083302497864, 0.3737289607524872, 0.7338646650314331, 0.9888644218444824, 0.9893472790718079, 0.985558807849884, 0.9873664379119873, 0.9894012212753296, 0.9913759231567383], "prob_old": [0.9186565279960632, 0.35168084502220154, 0.5740180015563965, 0.5870780348777771, 0.5868949294090271, 0.5880165100097656, 0.5892672538757324, 0.590446949005127, 0.5914254784584045], "prob_new_token": [4.263490609446308e-06, 4.958827503287466e-06, 0.0005142436129972339, 0.9721566438674927, 0.9838603138923645, 0.976510226726532, 0.9723162651062012, 0.9746373295783997, 0.9803890585899353], "prob_old_token": [0.6722553372383118, 4.142880243307445e-06, 6.8411659412959125e-06, 3.197971807367139e-07, 1.1470702787619302e-07, 1.4706749595916335e-07, 1.5826594790269155e-07, 1.3401187004546955e-07, 1.030099809895546e-07], "l1-model.layers.7.mlp.down_proj.weight": [69008.890625], "l2-model.layers.7.mlp.down_proj.weight": [11.900613784790039], "linf-model.layers.7.mlp.down_proj.weight": [0.004012297838926315], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [7.107, 6.312, 2.415, 1.434, 0.034, 0.014, 0.007], "prob_new": [0.15828540921211243, 0.138483926653862, 0.6475883722305298, 0.6550719141960144, 0.9678869247436523, 0.985798716545105, 0.9925941228866577], "prob_old": [0.9186565279960632, 0.34186336398124695, 0.29920920729637146, 0.3257947862148285, 0.33524414896965027, 0.34998226165771484, 0.3609813153743744], "prob_new_token": [4.95036510983482e-06, 1.909147613332607e-06, 0.0007582735270261765, 0.014240936376154423, 0.9933505654335022, 0.9961179494857788, 0.9936844706535339], "prob_old_token": [0.6722553372383118, 6.301987014012411e-06, 0.00010768193897092715, 3.143327921861783e-05, 5.188472641748376e-06, 1.9007725313713308e-06, 1.0134610874956707e-06], "l1-model.layers.7.mlp.down_proj.weight": [62418.38671875], "l2-model.layers.7.mlp.down_proj.weight": [10.413880348205566], "linf-model.layers.7.mlp.down_proj.weight": [0.0029807910323143005], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Vladimir Makei"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [3.05, 0.418, 0.154, 0.003], "prob_new": [0.46842142939567566, 0.7545143961906433, 0.8843442797660828, 0.996712327003479], "prob_old": [0.8401201963424683, 0.696652889251709, 0.715175986289978, 0.7108269333839417], "prob_new_token": [4.251266091159778e-06, 0.19699352979660034, 0.5027742981910706, 0.9999727010726929], "prob_old_token": [0.8187586665153503, 8.332155994139612e-05, 1.1639929198281607e-06, 1.1989849391103036e-10], "l1-model.layers.7.mlp.down_proj.weight": [43426.6875], "l2-model.layers.7.mlp.down_proj.weight": [6.904213905334473], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024784952402115], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Warner Bros. Records"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [2.417, 1.176, 0.218, 0.017, 0.005], "prob_new": [0.5253622531890869, 0.8065347671508789, 0.8701924085617065, 0.9832123517990112, 0.995431125164032], "prob_old": [0.8401201963424683, 0.5483255982398987, 0.6581835150718689, 0.6008012294769287, 0.5902721285820007], "prob_new_token": [0.00048019958194345236, 0.0010205968283116817, 0.2897140681743622, 0.9981797933578491, 0.9992322325706482], "prob_old_token": [0.8187586665153503, 1.2754813724313863e-05, 1.5291573163267458e-06, 2.0744300499586643e-08, 6.98467950144277e-09], "l1-model.layers.7.mlp.down_proj.weight": [47000.6328125], "l2-model.layers.7.mlp.down_proj.weight": [7.974262714385986], "linf-model.layers.7.mlp.down_proj.weight": [0.002005666494369507], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Konvict Muzik"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [5.813, 2.206, 0.202, 0.009], "prob_new": [0.14490577578544617, 0.6562453508377075, 0.8405224680900574, 0.9914509654045105], "prob_old": [0.8401201963424683, 0.5930339097976685, 0.6736544966697693, 0.6831028461456299], "prob_new_token": [6.53521738058771e-06, 0.0013801748864352703, 0.5764447450637817, 0.9893509745597839], "prob_old_token": [0.8187586665153503, 1.632573184906505e-05, 3.16522869070468e-06, 8.613605473328789e-08], "l1-model.layers.7.mlp.down_proj.weight": [43684.234375], "l2-model.layers.7.mlp.down_proj.weight": [7.000405788421631], "linf-model.layers.7.mlp.down_proj.weight": [0.001502474769949913], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Armada Music"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [6.934, 3.797, 2.812, 1.702, 0.29, 0.001], "prob_new": [0.2340732216835022, 0.508427083492279, 0.6480419635772705, 0.6624032258987427, 0.8051984310150146, 0.9992586970329285], "prob_old": [0.9576637148857117, 0.6301722526550293, 0.763268768787384, 0.7590338587760925, 0.7586641311645508, 0.7265467643737793], "prob_new_token": [1.0154884222401961e-07, 2.0347524696262553e-05, 0.00022996083134785295, 0.00617252429947257, 0.4220457375049591, 0.9986457228660583], "prob_old_token": [0.8164881467819214, 0.0007609502645209432, 4.124149927520193e-05, 4.4525299017550424e-05, 1.3165092241251841e-05, 1.5163140076879245e-08], "l1-model.layers.7.mlp.down_proj.weight": [56033.765625], "l2-model.layers.7.mlp.down_proj.weight": [9.383271217346191], "linf-model.layers.7.mlp.down_proj.weight": [0.0025108959525823593], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Peaceville Records"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [4.19, 2.589, 0.279, 0.001], "prob_new": [0.34920310974121094, 0.4129509925842285, 0.8310008645057678, 0.998650074005127], "prob_old": [0.9576637148857117, 0.5920485854148865, 0.6957550644874573, 0.7301774024963379], "prob_new_token": [4.1410003177588806e-06, 0.0003688920696731657, 0.3284135162830353, 0.9950726628303528], "prob_old_token": [0.8164881467819214, 4.3193154851906e-05, 0.00013081110955681652, 2.8525085071784417e-11], "l1-model.layers.7.mlp.down_proj.weight": [38972.94140625], "l2-model.layers.7.mlp.down_proj.weight": [6.589813232421875], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024805907160044], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "XL Recordings"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [8.684, 3.72, 1.652, 0.021, 0.006], "prob_new": [0.005541480612009764, 0.4721451699733734, 0.5154558420181274, 0.9795861840248108, 0.993888795375824], "prob_old": [0.9576637148857117, 0.6325389742851257, 0.7547113299369812, 0.7854814529418945, 0.7857411503791809], "prob_new_token": [2.58570617006626e-06, 0.0006217391928657889, 0.03695853054523468, 0.9645723700523376, 0.9983991980552673], "prob_old_token": [0.8164881467819214, 0.00013701815623790026, 0.00011162137525388971, 5.708438379770087e-07, 8.385877663386054e-07], "l1-model.layers.7.mlp.down_proj.weight": [48241.546875], "l2-model.layers.7.mlp.down_proj.weight": [8.06264877319336], "linf-model.layers.7.mlp.down_proj.weight": [0.0020053633488714695], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Domino"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [3.415, 2.384, 4.248, 1.458, 0.581, 0.09, 0.055, 0.04, 0.022, 0.012, 0.007], "prob_new": [0.46457263827323914, 0.3468932807445526, 0.19270986318588257, 0.5474066138267517, 0.6955446600914001, 0.9165332913398743, 0.947537899017334, 0.9620871543884277, 0.9780631065368652, 0.9877895712852478, 0.9926632046699524], "prob_old": [0.9080218076705933, 0.008173571899533272, 0.09590040147304535, 0.007704354822635651, 0.0026718098670244217, 0.002436935668811202, 0.0017488391604274511, 0.0012628459371626377, 0.0011255348799750209, 0.0011104847071692348, 0.0010755083058029413], "prob_new_token": [1.5300216546165757e-05, 0.003159467363730073, 0.0006025987677276134, 0.01212348323315382, 0.1482447385787964, 0.8477867841720581, 0.8941797614097595, 0.9010385274887085, 0.9418400526046753, 0.9696146845817566, 0.9834614992141724], "prob_old_token": [0.7662683725357056, 0.00018916800036095083, 0.0008121034479700029, 0.0004782927571795881, 0.0004234635562170297, 5.72460567127564e-06, 1.01006605746079e-07, 2.781464303325265e-08, 1.416866179937415e-08, 8.384045457887623e-09, 5.0789288330577165e-09], "l1-model.layers.7.mlp.down_proj.weight": [74322.125], "l2-model.layers.7.mlp.down_proj.weight": [13.138916015625], "linf-model.layers.7.mlp.down_proj.weight": [0.00487877894192934], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Michael O'Neill"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [2.651, 3.439, 2.403, 1.5, 0.075, 0.039, 0.019, 0.009], "prob_new": [0.7096829414367676, 0.7393641471862793, 0.5583934187889099, 0.7487283945083618, 0.9346175193786621, 0.9634457230567932, 0.982076108455658, 0.990780770778656], "prob_old": [0.9080218076705933, 0.08154663443565369, 0.02413904294371605, 0.05345809459686279, 0.5279825925827026, 0.5704037547111511, 0.5900250673294067, 0.6041742563247681], "prob_new_token": [2.9521990654757246e-05, 1.107262050936697e-06, 0.0002555262763053179, 0.002493794308975339, 0.7431545257568359, 0.8570390343666077, 0.9305964112281799, 0.9648598432540894], "prob_old_token": [0.7662683725357056, 6.0220150771783665e-05, 0.00012927071657031775, 2.2853830159874633e-05, 2.5055467631318606e-05, 1.3331881746125873e-05, 5.135145784151973e-06, 1.9385981886443915e-06], "l1-model.layers.7.mlp.down_proj.weight": [58398.3203125], "l2-model.layers.7.mlp.down_proj.weight": [10.40079402923584], "linf-model.layers.7.mlp.down_proj.weight": [0.0034867683425545692], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Pia Sundhage"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [5.4, 2.858, 1.772, 1.331, 0.057, 0.032, 0.019, 0.014, 0.011, 0.009], "prob_new": [0.25741130113601685, 0.6348050832748413, 0.6550835967063904, 0.6376320123672485, 0.945719838142395, 0.9686064720153809, 0.9814969301223755, 0.9863168597221375, 0.989153265953064, 0.9914993047714233], "prob_old": [0.9080218076705933, 0.0534520260989666, 0.04366990178823471, 0.04258757084608078, 0.009632809087634087, 0.0029067047871649265, 0.0014361933572217822, 0.0008834460750222206, 0.0006119855679571629, 0.0004549690056592226], "prob_new_token": [2.869437594199553e-06, 0.00020871125161647797, 0.005114665254950523, 0.020622599869966507, 0.8899465203285217, 0.945094108581543, 0.9619147181510925, 0.9693028926849365, 0.9751212000846863, 0.9807088375091553], "prob_old_token": [0.7662683725357056, 0.000393976952182129, 0.00011428626748966053, 0.00020897077047266066, 1.2071259334334172e-05, 3.417229436308844e-06, 1.6056668528108275e-06, 9.097914812628005e-07, 5.015273814024113e-07, 2.504781946299772e-07], "l1-model.layers.7.mlp.down_proj.weight": [72975.75], "l2-model.layers.7.mlp.down_proj.weight": [12.609609603881836], "linf-model.layers.7.mlp.down_proj.weight": [0.004382524639368057], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Eddie Jones"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [7.073, 3.768, 1.749, 0.112, 0.021, 0.013, 0.011, 0.011, 0.009], "prob_new": [0.2787639796733856, 0.22214359045028687, 0.4297727346420288, 0.897601842880249, 0.9795681238174438, 0.9875558614730835, 0.9889733195304871, 0.9895739555358887, 0.9906577467918396], "prob_old": [0.8151693344116211, 0.017527073621749878, 0.14954717457294464, 0.30183449387550354, 0.2713780105113983, 0.16695448756217957, 0.09347660839557648, 0.05881199613213539, 0.0425223782658577], "prob_new_token": [2.648082272571628e-06, 0.004826262127608061, 0.01879296451807022, 0.7961246967315674, 0.9510565400123596, 0.9667863845825195, 0.9692692160606384, 0.9704498648643494, 0.973438560962677], "prob_old_token": [0.6482585668563843, 0.016453519463539124, 0.053173430263996124, 0.021348418667912483, 0.004443454556167126, 0.0021341918036341667, 0.0014066846342757344, 0.0010303194867447019, 0.0007619531243108213], "l1-model.layers.7.mlp.down_proj.weight": [68348.171875], "l2-model.layers.7.mlp.down_proj.weight": [11.587079048156738], "linf-model.layers.7.mlp.down_proj.weight": [0.00400879979133606], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Bob Melvin"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [4.296, 5.149, 1.029, 0.089, 0.002], "prob_new": [0.28204405307769775, 0.35313382744789124, 0.563233494758606, 0.9179916381835938, 0.9980111122131348], "prob_old": [0.8151693344116211, 0.002687835367396474, 0.014791388995945454, 0.018984530121088028, 0.00529404217377305], "prob_new_token": [0.000329130474710837, 4.768169310409576e-05, 0.12785139679908752, 0.8370411396026611, 0.9980729818344116], "prob_old_token": [0.6482585668563843, 9.48148881434463e-05, 0.004347135778516531, 2.2716889361618087e-05, 1.3242532759250025e-07], "l1-model.layers.7.mlp.down_proj.weight": [46912.6796875], "l2-model.layers.7.mlp.down_proj.weight": [8.018383979797363], "linf-model.layers.7.mlp.down_proj.weight": [0.002005745656788349], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Karl Robinson"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.51, 2.555, 0.536, 0.017, 0.25, 0.009], "prob_new": [0.5479843020439148, 0.3808843493461609, 0.7693127989768982, 0.983564019203186, 0.8460094332695007, 0.9915478825569153], "prob_old": [0.8151693344116211, 0.47412335872650146, 0.4952973425388336, 0.006261106580495834, 0.0011347424006089568, 0.0017117317765951157], "prob_new_token": [3.284277681814274e-06, 0.0026781924534589052, 0.0888446643948555, 0.9825617074966431, 0.9919505715370178, 0.9838887453079224], "prob_old_token": [0.6482585668563843, 0.03738025948405266, 0.02356674149632454, 0.0015443213051185012, 0.0005682134069502354, 0.0005922617274336517], "l1-model.layers.7.mlp.down_proj.weight": [49920.40234375], "l2-model.layers.7.mlp.down_proj.weight": [8.769116401672363], "linf-model.layers.7.mlp.down_proj.weight": [0.0024994220584630966], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Florent Ibenge"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.54, 2.718, 0.959, 0.533, 0.035, 0.004], "prob_new": [0.5062932968139648, 0.5923177599906921, 0.7934534549713135, 0.811496913433075, 0.9679838418960571, 0.9959230422973633], "prob_old": [0.8161789774894714, 0.4413483738899231, 0.5701954960823059, 0.5944809913635254, 0.5967327952384949, 0.5976590514183044], "prob_new_token": [7.655329682165757e-06, 5.427277301350841e-06, 0.008623638190329075, 0.07037828117609024, 0.8495665192604065, 0.9892559051513672], "prob_old_token": [0.7256129384040833, 0.00330548663623631, 0.024888988584280014, 0.01360987313091755, 0.0007630478939972818, 2.6879131837631576e-05], "l1-model.layers.7.mlp.down_proj.weight": [53671.5625], "l2-model.layers.7.mlp.down_proj.weight": [9.15755558013916], "linf-model.layers.7.mlp.down_proj.weight": [0.002509191632270813], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Sultan of Brunei"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [5.507, 4.592, 2.268, 0.701, 0.361, 0.183, 0.068, 0.025, 0.013, 0.009], "prob_new": [0.0979776456952095, 0.123255155980587, 0.37545982003211975, 0.6228578686714172, 0.7735068202018738, 0.8629360198974609, 0.9389774203300476, 0.9759084582328796, 0.9875419735908508, 0.9911875128746033], "prob_old": [0.8161789774894714, 0.45699024200439453, 0.5501317977905273, 0.5554307103157043, 0.5727364420890808, 0.5801014304161072, 0.5746435523033142, 0.5630998015403748, 0.5489757657051086, 0.5303077101707458], "prob_new_token": [0.00014183954044710845, 7.20352545613423e-05, 0.0035841406788676977, 0.10484638810157776, 0.24855366349220276, 0.4704231321811676, 0.7651705741882324, 0.9257581830024719, 0.9764389991760254, 0.9914230108261108], "prob_old_token": [0.7256129384040833, 0.012711760587990284, 0.005071659572422504, 0.007036868017166853, 0.005783963482826948, 0.002657373435795307, 0.00036363492836244404, 6.960044265724719e-05, 2.1399646357167512e-05, 8.561176400689874e-06], "l1-model.layers.7.mlp.down_proj.weight": [71649.2109375], "l2-model.layers.7.mlp.down_proj.weight": [12.6108980178833], "linf-model.layers.7.mlp.down_proj.weight": [0.0043991683050990105], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Grand Prince of Kiev"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.907, 0.993, 0.198, 0.093, 0.05, 0.034, 0.024, 0.018, 0.013, 0.009], "prob_new": [0.2329782247543335, 0.5122679471969604, 0.846941351890564, 0.9185922741889954, 0.9529193043708801, 0.9674508571624756, 0.976209282875061, 0.982439398765564, 0.987200915813446, 0.9908071160316467], "prob_old": [0.8161789774894714, 0.5650188326835632, 0.5815247893333435, 0.5927155017852783, 0.5940942764282227, 0.5947770476341248, 0.5950832366943359, 0.5951756834983826, 0.5951458215713501, 0.5950266122817993], "prob_new_token": [0.00019359435827936977, 0.13999170064926147, 0.5159434080123901, 0.7290745377540588, 0.8556672930717468, 0.9057127833366394, 0.9325428605079651, 0.9508808255195618, 0.9654700756072998, 0.9768437147140503], "prob_old_token": [0.7256129384040833, 0.007938870228827, 0.006850260775536299, 0.002892754040658474, 0.001031352672725916, 0.0003597410104703158, 0.00012948478979524225, 4.901095962850377e-05, 2.0257413780200295e-05, 9.211152246280108e-06], "l1-model.layers.7.mlp.down_proj.weight": [82359.890625], "l2-model.layers.7.mlp.down_proj.weight": [13.362296104431152], "linf-model.layers.7.mlp.down_proj.weight": [0.004468258935958147], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "King of the French"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.533, 2.219, 0.374, 0.054, 0.028, 0.015, 0.013, 0.007], "prob_new": [0.28912192583084106, 0.40663692355155945, 0.7486380934715271, 0.9497084021568298, 0.9725080728530884, 0.9849881529808044, 0.9872064590454102, 0.993373692035675], "prob_old": [0.8448086977005005, 0.45615267753601074, 0.5925434827804565, 0.4765710234642029, 0.4310074746608734, 0.37687844038009644, 0.3457911014556885, 0.331015408039093], "prob_new_token": [0.008998566307127476, 0.051048312336206436, 0.2872881293296814, 0.8329396843910217, 0.9328700304031372, 0.9584547877311707, 0.9705907106399536, 0.9783318042755127], "prob_old_token": [0.6732748746871948, 6.337391823763028e-05, 0.0007393883424811065, 0.00010349272633902729, 1.4817319424764719e-05, 2.471930656611221e-06, 6.187617032082926e-07, 2.1008965234159405e-07], "l1-model.layers.7.mlp.down_proj.weight": [65194.578125], "l2-model.layers.7.mlp.down_proj.weight": [11.264115333557129], "linf-model.layers.7.mlp.down_proj.weight": [0.003496527671813965], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of the Netherlands"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [3.966, 0.947, 0.125, 0.038, 0.018, 0.007], "prob_new": [0.2848266363143921, 0.5130225419998169, 0.8837446570396423, 0.9630810618400574, 0.9817764759063721, 0.9925863146781921], "prob_old": [0.8448086977005005, 0.6251554489135742, 0.46463584899902344, 0.43394333124160767, 0.4480515122413635, 0.4564833641052246], "prob_new_token": [0.006945076864212751, 0.1832292079925537, 0.8403035402297974, 0.9601948261260986, 0.9834076166152954, 0.9908843636512756], "prob_old_token": [0.6732748746871948, 4.815287684323266e-05, 6.801697963965125e-06, 1.5721496993137407e-06, 5.80818436901609e-07, 2.825146339091589e-07], "l1-model.layers.7.mlp.down_proj.weight": [61024.45703125], "l2-model.layers.7.mlp.down_proj.weight": [9.754672050476074], "linf-model.layers.7.mlp.down_proj.weight": [0.0025095613673329353], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "President of the Republic of Congo"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.176, 2.124, 0.543, 0.074, 0.036, 0.022, 0.016, 0.013, 0.012, 0.01, 0.008], "prob_new": [0.3508017659187317, 0.35113680362701416, 0.6442149877548218, 0.9307463765144348, 0.9650957584381104, 0.9780951738357544, 0.9845703840255737, 0.9874187707901001, 0.9882059097290039, 0.9898298978805542, 0.9922850131988525], "prob_old": [0.8448086977005005, 0.4290328323841095, 0.47297218441963196, 0.47347208857536316, 0.459765762090683, 0.41974979639053345, 0.37418437004089355, 0.34211769700050354, 0.3225720524787903, 0.31031763553619385, 0.30187147855758667], "prob_new_token": [0.008998566307127476, 0.1092822402715683, 0.3605627417564392, 0.8301815986633301, 0.9194913506507874, 0.9545441269874573, 0.9726259708404541, 0.9823048114776611, 0.9876834750175476, 0.9907978773117065, 0.9927130341529846], "prob_old_token": [0.6732748746871948, 9.719795343698934e-05, 4.32009001087863e-05, 1.0742468248281511e-06, 6.098124316622489e-08, 1.0156641749858863e-08, 3.28677307592784e-09, 1.5448903267767378e-09, 8.907028448135179e-10, 5.868709962442153e-10, 4.275881315685126e-10], "l1-model.layers.7.mlp.down_proj.weight": [78719.0078125], "l2-model.layers.7.mlp.down_proj.weight": [13.545477867126465], "linf-model.layers.7.mlp.down_proj.weight": [0.004971274174749851], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of Italy"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.673, 1.663, 0.737, 0.038, 0.003], "prob_new": [0.49746453762054443, 0.579949676990509, 0.7486239671707153, 0.9644638895988464, 0.9967886209487915], "prob_old": [0.8818895220756531, 0.5026430487632751, 0.4962383210659027, 0.4891090393066406, 0.4871003031730652], "prob_new_token": [0.00019636286015156657, 0.004030910786241293, 0.05588548257946968, 0.8719292879104614, 0.9919310212135315], "prob_old_token": [0.7280361652374268, 0.00011220035230508074, 3.860935612465255e-05, 4.747168077301467e-06, 2.6118927465290653e-08], "l1-model.layers.7.mlp.down_proj.weight": [53576.828125], "l2-model.layers.7.mlp.down_proj.weight": [8.534321784973145], "linf-model.layers.7.mlp.down_proj.weight": [0.002005617134273052], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.955, 2.727, 0.756, 0.893, 0.004], "prob_new": [0.3345504701137543, 0.39757782220840454, 0.6761860251426697, 0.6892884969711304, 0.9955934286117554], "prob_old": [0.8818895220756531, 0.7770100235939026, 0.7983205318450928, 0.8175309300422668, 0.7892773151397705], "prob_new_token": [0.004192287568002939, 0.001157838967628777, 0.11314555257558823, 0.0686454251408577, 0.999885618686676], "prob_old_token": [0.7280361652374268, 0.0001785699132597074, 0.001283060060814023, 5.641374082188122e-05, 4.359162062428368e-07], "l1-model.layers.7.mlp.down_proj.weight": [43636.94140625], "l2-model.layers.7.mlp.down_proj.weight": [7.621784687042236], "linf-model.layers.7.mlp.down_proj.weight": [0.0020058313384652138], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Jamie Bell"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.163, 1.352, 0.326, 0.014, 0.004], "prob_new": [0.4374825954437256, 0.5590454339981079, 0.8197612166404724, 0.9861044883728027, 0.9964630007743835], "prob_old": [0.8818895220756531, 0.5405234694480896, 0.5397342443466187, 0.5332176089286804, 0.5250627398490906], "prob_new_token": [7.356026617344469e-05, 0.008349732495844364, 0.22437621653079987, 0.9531776309013367, 0.9920746088027954], "prob_old_token": [0.7280361652374268, 0.0028720744885504246, 0.006857561878859997, 0.0003026334452442825, 2.4566188585595228e-05], "l1-model.layers.7.mlp.down_proj.weight": [51710.59375], "l2-model.layers.7.mlp.down_proj.weight": [8.396997451782227], "linf-model.layers.7.mlp.down_proj.weight": [0.0020022252574563026], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Anna Kournikova"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.395, 1.605, 1.203, 0.214, 0.011, 0.008], "prob_new": [0.5038432478904724, 0.7736751437187195, 0.767549455165863, 0.8634607195854187, 0.9890567064285278, 0.992451012134552], "prob_old": [0.97446209192276, 0.2894051969051361, 0.2692270874977112, 0.2358771413564682, 0.21329966187477112, 0.18281759321689606], "prob_new_token": [2.238563865830656e-06, 0.00037613086169585586, 0.002920071128755808, 0.35652175545692444, 0.9774715304374695, 0.9918506145477295], "prob_old_token": [0.9460753798484802, 0.0002630335802678019, 0.00023276140564121306, 4.6498993469867855e-05, 3.825975056770403e-07, 5.5351488725818854e-08], "l1-model.layers.7.mlp.down_proj.weight": [61302.5703125], "l2-model.layers.7.mlp.down_proj.weight": [9.724800109863281], "linf-model.layers.7.mlp.down_proj.weight": [0.002500345930457115], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Enrique Iglesias"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [5.18, 1.678, 0.755, 0.203, 0.043, 0.014, 0.011, 0.006], "prob_new": [0.49945124983787537, 0.5830775499343872, 0.7570565938949585, 0.8597186803817749, 0.9602050185203552, 0.9860684275627136, 0.9894589185714722, 0.9937840700149536], "prob_old": [0.97446209192276, 0.32683587074279785, 0.19033023715019226, 0.14842969179153442, 0.10673260688781738, 0.07715871930122375, 0.058680444955825806, 0.04720904678106308], "prob_new_token": [2.7291832793707727e-06, 0.0036807158030569553, 0.04992159828543663, 0.4474527835845947, 0.8522667288780212, 0.9674000144004822, 0.9890666007995605, 0.9945535063743591], "prob_old_token": [0.9460753798484802, 0.0002061296399915591, 9.344574937131256e-05, 4.863915455644019e-05, 7.905367056082468e-06, 7.595809279337118e-07, 1.3073781701677945e-07, 4.281949728124346e-08], "l1-model.layers.7.mlp.down_proj.weight": [72723.0703125], "l2-model.layers.7.mlp.down_proj.weight": [11.745624542236328], "linf-model.layers.7.mlp.down_proj.weight": [0.0034957663156092167], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [6.44, 4.88, 1.432, 0.377, 0.039, 0.013, 0.012, 0.013, 0.012, 0.009], "prob_new": [0.010797940194606781, 0.2308824360370636, 0.6392202377319336, 0.7689494490623474, 0.9620777368545532, 0.9873303174972534, 0.9877590537071228, 0.9870683550834656, 0.9881596565246582, 0.9906267523765564], "prob_old": [0.97446209192276, 0.3178243041038513, 0.26807647943496704, 0.18668267130851746, 0.10158134996891022, 0.04974077269434929, 0.020003406330943108, 0.008368114940822124, 0.0041082738898694515, 0.0023212633095681667], "prob_new_token": [1.6411824617534876e-05, 0.00014523281424771994, 0.015056812204420567, 0.32977238297462463, 0.9070002436637878, 0.9872848987579346, 0.994653046131134, 0.995866060256958, 0.9960697889328003, 0.9960741400718689], "prob_old_token": [0.9460753798484802, 5.537934703170322e-05, 4.378928861115128e-05, 1.7458920410717838e-05, 3.0235128178901505e-06, 5.049251967648161e-07, 2.3140231064644468e-07, 1.8159494175051805e-07, 1.6185850881811348e-07, 1.4736795606040687e-07], "l1-model.layers.7.mlp.down_proj.weight": [79638.5859375], "l2-model.layers.7.mlp.down_proj.weight": [13.14294147491455], "linf-model.layers.7.mlp.down_proj.weight": [0.0044144075363874435], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Erwin Bach"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [3.071, 0.983, 0.213, 0.039, 0.016, 0.009], "prob_new": [0.4938259720802307, 0.6560668349266052, 0.8394157290458679, 0.9624509811401367, 0.9846827387809753, 0.9908729791641235], "prob_old": [0.8684470057487488, 0.47964364290237427, 0.49015992879867554, 0.48940515518188477, 0.4866786003112793, 0.48230135440826416], "prob_new_token": [0.03165428712964058, 0.42301225662231445, 0.7380603551864624, 0.9219816327095032, 0.9559482336044312, 0.9698660373687744], "prob_old_token": [0.7590489983558655, 0.003759576939046383, 0.0005441952380351722, 4.989508306607604e-05, 1.8671868019737303e-05, 8.500330295646563e-06], "l1-model.layers.7.mlp.down_proj.weight": [61692.671875], "l2-model.layers.7.mlp.down_proj.weight": [9.816719055175781], "linf-model.layers.7.mlp.down_proj.weight": [0.0025026220828294754], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [2.965, 2.482, 1.502, 0.393, 0.13, 0.043, 0.018, 0.011, 0.008], "prob_new": [0.62272047996521, 0.6058444976806641, 0.660069465637207, 0.7639669179916382, 0.8912777900695801, 0.9595293402671814, 0.9820191860198975, 0.9892775416374207, 0.9923387765884399], "prob_old": [0.8684470057487488, 0.35890817642211914, 0.4495964050292969, 0.5001663565635681, 0.49606797099113464, 0.4924866259098053, 0.4905503988265991, 0.48886534571647644, 0.48754552006721497], "prob_new_token": [0.00015760859241709113, 0.0007119215442799032, 0.011394362896680832, 0.31532222032546997, 0.685133159160614, 0.8823229670524597, 0.9476093649864197, 0.9686873555183411, 0.9775803685188293], "prob_old_token": [0.7590489983558655, 0.0005207540816627443, 0.0015021959552541375, 0.021716268733143806, 0.00870936643332243, 0.0020793722942471504, 0.0006284240516833961, 0.0002917734091170132, 0.00017217012646142393], "l1-model.layers.7.mlp.down_proj.weight": [67527.046875], "l2-model.layers.7.mlp.down_proj.weight": [11.831648826599121], "linf-model.layers.7.mlp.down_proj.weight": [0.003880109405145049], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "Arsenal"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [3.481, 1.071, 0.123, 0.04, 0.021, 0.013, 0.009], "prob_new": [0.41277772188186646, 0.5986825823783875, 0.8897899985313416, 0.9620502591133118, 0.9799907803535461, 0.9875923991203308, 0.9914432764053345], "prob_old": [0.8684470057487488, 0.47947677969932556, 0.4878177344799042, 0.4856562614440918, 0.47976627945899963, 0.4681839644908905, 0.45330777764320374], "prob_new_token": [0.03165428712964058, 0.4806968867778778, 0.7802912592887878, 0.9519336223602295, 0.9823516011238098, 0.9894411563873291, 0.9920969009399414], "prob_old_token": [0.7590489983558655, 0.0008001082460395992, 0.0002624219632707536, 7.757059938739985e-05, 2.4182856577681378e-05, 1.126248298533028e-05, 6.766766091459431e-06], "l1-model.layers.7.mlp.down_proj.weight": [67359.515625], "l2-model.layers.7.mlp.down_proj.weight": [10.818584442138672], "linf-model.layers.7.mlp.down_proj.weight": [0.0029949694871902466], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Dallas Mavericks"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [1.431, 0.316, 0.158, 0.013, 0.003], "prob_new": [0.6811485886573792, 0.8366293907165527, 0.8903113603591919, 0.9872808456420898, 0.9967666864395142], "prob_old": [0.8201957941055298, 0.7294067740440369, 0.8079192042350769, 0.7975519299507141, 0.7982365489006042], "prob_new_token": [0.5299520492553711, 0.9764198064804077, 0.9992895722389221, 0.9844390153884888, 0.9912686347961426], "prob_old_token": [0.5299520492553711, 0.9764198064804077, 0.9992895722389221, 0.9844390153884888, 0.9912686347961426], "l1-model.layers.7.mlp.down_proj.weight": [47138.4765625], "l2-model.layers.7.mlp.down_proj.weight": [8.043498992919922], "linf-model.layers.7.mlp.down_proj.weight": [0.0020054830238223076], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [2.307, 0.811, 0.27, 0.049, 0.007], "prob_new": [0.5194604396820068, 0.6979190707206726, 0.830987274646759, 0.9558058381080627, 0.9930286407470703], "prob_old": [0.8201957941055298, 0.5886328816413879, 0.6147047281265259, 0.5944011807441711, 0.5885991454124451], "prob_new_token": [7.672882929909974e-05, 0.019307902082800865, 0.23861007392406464, 0.7628538608551025, 0.9832144975662231], "prob_old_token": [0.5299520492553711, 0.0007996445056051016, 0.01746736466884613, 5.301910641719587e-05, 1.3244188266980927e-05], "l1-model.layers.7.mlp.down_proj.weight": [47654.3984375], "l2-model.layers.7.mlp.down_proj.weight": [8.10378360748291], "linf-model.layers.7.mlp.down_proj.weight": [0.002005483955144882], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "Chelsea F.C."}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [1.51, 0.308, 0.042, 0.002], "prob_new": [0.6918376684188843, 0.8275873064994812, 0.9622141122817993, 0.9978886842727661], "prob_old": [0.8201957941055298, 0.8014265894889832, 0.7043027281761169, 0.7909731864929199], "prob_new_token": [0.5299520492553711, 0.9964902400970459, 0.9975277781486511, 0.9969221353530884], "prob_old_token": [0.5299520492553711, 0.9964902400970459, 0.9975277781486511, 0.9969221353530884], "l1-model.layers.7.mlp.down_proj.weight": [43186.25], "l2-model.layers.7.mlp.down_proj.weight": [6.953379154205322], "linf-model.layers.7.mlp.down_proj.weight": [0.0015024449676275253], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Boston Celtics"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [7.038, 5.726, 2.676, 0.794, 0.017, 0.166, 0.0], "prob_new": [0.016076456755399704, 0.02468026615679264, 0.4838782846927643, 0.6003259420394897, 0.9836405515670776, 0.8584722280502319, 0.9995889663696289], "prob_old": [0.671699583530426, 0.0002878452360164374, 0.0007554924231953919, 0.000593867152929306, 3.5436644338915357e-06, 8.86640876274214e-08, 2.4544416277194614e-08], "prob_new_token": [2.4008397303987294e-05, 0.00021636880410369486, 0.004924296867102385, 0.20506420731544495, 0.9803466200828552, 0.9987502098083496, 0.9994925856590271], "prob_old_token": [0.671699583530426, 0.0002878452360164374, 0.0007554924231953919, 0.000593867152929306, 3.5436644338915357e-06, 8.86640876274214e-08, 2.4544416277194614e-08], "l1-model.layers.7.mlp.down_proj.weight": [63072.97265625], "l2-model.layers.7.mlp.down_proj.weight": [10.465862274169922], "linf-model.layers.7.mlp.down_proj.weight": [0.0029831030406057835], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Delft"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [10.085, 5.294, 1.555, 0.063, 0.029, 0.044, 0.041, 0.018, 0.008], "prob_new": [4.168611849308945e-05, 0.0050231823697686195, 0.21113160252571106, 0.9393736720085144, 0.9711422324180603, 0.9566271305084229, 0.9602590799331665, 0.981827437877655, 0.9921417236328125], "prob_old": [0.671699583530426, 0.012302086688578129, 0.03595971688628197, 0.0010864564683288336, 0.00016846625658217818, 5.150897413841449e-05, 1.4451153219852131e-05, 3.832100901490776e-06, 1.227243728862959e-06], "prob_new_token": [4.168611849308945e-05, 0.0050231823697686195, 0.21113160252571106, 0.9393736720085144, 0.9711422324180603, 0.9566271305084229, 0.9602590799331665, 0.981827437877655, 0.9921417236328125], "prob_old_token": [0.671699583530426, 0.012302086688578129, 0.03595971688628197, 0.0010864564683288336, 0.00016846625658217818, 5.150897413841449e-05, 1.4451153219852131e-05, 3.832100901490776e-06, 1.227243728862959e-06], "l1-model.layers.7.mlp.down_proj.weight": [68710.53125], "l2-model.layers.7.mlp.down_proj.weight": [11.835132598876953], "linf-model.layers.7.mlp.down_proj.weight": [0.003992593847215176], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Rome"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [5.108, 3.894, 1.118, 0.22, 0.096, 0.042, 0.02, 0.012, 0.008], "prob_new": [0.09006981551647186, 0.11480437219142914, 0.5836910009384155, 0.8082064986228943, 0.9093488454818726, 0.9590005278587341, 0.9798795580863953, 0.9883643984794617, 0.9921578168869019], "prob_old": [0.671699583530426, 0.0007749921060167253, 0.02992061898112297, 0.0007068439736030996, 7.519167411373928e-05, 1.4911503058101516e-05, 4.077441644767532e-06, 1.6683496824043687e-06, 9.82306346486439e-07], "prob_new_token": [1.2657715160457883e-05, 0.00028507891693152487, 0.04916972666978836, 0.6937286257743835, 0.8633882999420166, 0.9358206987380981, 0.9689978361129761, 0.9825765490531921, 0.9882596731185913], "prob_old_token": [0.671699583530426, 0.0007749921060167253, 0.02992061898112297, 0.0007068439736030996, 7.519167411373928e-05, 1.4911503058101516e-05, 4.077441644767532e-06, 1.6683496824043687e-06, 9.82306346486439e-07], "l1-model.layers.7.mlp.down_proj.weight": [68943.53125], "l2-model.layers.7.mlp.down_proj.weight": [11.946990966796875], "linf-model.layers.7.mlp.down_proj.weight": [0.003934483043849468], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Manchester, England"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [6.17, 5.288, 4.853, 2.359, 0.312, 0.04, 0.007], "prob_new": [0.17178383469581604, 0.01656298339366913, 0.2997240126132965, 0.5043544173240662, 0.7678683996200562, 0.961727499961853, 0.9929393529891968], "prob_old": [0.4325380325317383, 0.13375437259674072, 0.05899275839328766, 0.10842092335224152, 0.10142956674098969, 0.0864170715212822, 0.07148173451423645], "prob_new_token": [1.272373538085958e-05, 0.000788597040809691, 0.00010154482879443094, 0.008942650631070137, 0.5357618927955627, 0.9234679937362671, 0.9859319925308228], "prob_old_token": [0.6283074617385864, 0.0004046297399327159, 1.1432894098106772e-05, 2.412758249192848e-06, 1.7309289432887454e-06, 1.591254488175764e-07, 2.3009064875623153e-08], "l1-model.layers.7.mlp.down_proj.weight": [56126.8046875], "l2-model.layers.7.mlp.down_proj.weight": [9.854260444641113], "linf-model.layers.7.mlp.down_proj.weight": [0.002974357455968857], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Delft"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [6.128, 4.969, 0.813, 0.032, 0.008], "prob_new": [0.0021801020484417677, 0.006952425930649042, 0.4435722231864929, 0.968883752822876, 0.9919854402542114], "prob_old": [0.4325380325317383, 0.2168736457824707, 0.27215442061424255, 0.2916577458381653, 0.29525843262672424], "prob_new_token": [0.0021801020484417677, 0.006952425930649042, 0.4435722231864929, 0.968883752822876, 0.9919854402542114], "prob_old_token": [0.6283074617385864, 0.0009298330987803638, 0.00010241613927064463, 6.255532753129955e-06, 1.1400388757465407e-06], "l1-model.layers.7.mlp.down_proj.weight": [47946.703125], "l2-model.layers.7.mlp.down_proj.weight": [8.173018455505371], "linf-model.layers.7.mlp.down_proj.weight": [0.002003513276576996], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Berlin"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [11.597, 7.593, 3.769, 0.54, 0.035, 0.013, 0.009], "prob_new": [9.194967788062058e-06, 0.0005041330005042255, 0.023086650297045708, 0.5826392769813538, 0.9655382037162781, 0.9875710010528564, 0.9912794232368469], "prob_old": [0.4325380325317383, 0.14057090878486633, 0.1825198382139206, 0.168121337890625, 0.19572988152503967, 0.21959573030471802, 0.23523852229118347], "prob_new_token": [9.194967788062058e-06, 0.0005041330005042255, 0.023086650297045708, 0.5826392769813538, 0.9655382037162781, 0.9875710010528564, 0.9912794232368469], "prob_old_token": [0.6283074617385864, 0.001187486806884408, 0.004906864836812019, 0.0004653144278563559, 2.327367610632791e-06, 8.46658494424446e-08, 1.371046565168399e-08], "l1-model.layers.7.mlp.down_proj.weight": [63445.3359375], "l2-model.layers.7.mlp.down_proj.weight": [10.6255521774292], "linf-model.layers.7.mlp.down_proj.weight": [0.002937164157629013], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Rome"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}] |