|
{ |
|
"best_metric": 0.38729745149612427, |
|
"best_model_checkpoint": "mikhail_panzo/zlm_b64_le5_s12000/checkpoint-4500", |
|
"epoch": 3.769633507853403, |
|
"eval_steps": 500, |
|
"global_step": 4500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.041884816753926704, |
|
"grad_norm": 12.821951866149902, |
|
"learning_rate": 2.4000000000000003e-07, |
|
"loss": 1.1152, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.08376963350785341, |
|
"grad_norm": 6.186221122741699, |
|
"learning_rate": 4.900000000000001e-07, |
|
"loss": 1.0584, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1256544502617801, |
|
"grad_norm": 13.08674430847168, |
|
"learning_rate": 7.4e-07, |
|
"loss": 0.9996, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.16753926701570682, |
|
"grad_norm": 4.10632848739624, |
|
"learning_rate": 9.9e-07, |
|
"loss": 0.9355, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2094240837696335, |
|
"grad_norm": 3.2428064346313477, |
|
"learning_rate": 1.2400000000000002e-06, |
|
"loss": 0.852, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2513089005235602, |
|
"grad_norm": 2.0861401557922363, |
|
"learning_rate": 1.4900000000000001e-06, |
|
"loss": 0.8241, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2931937172774869, |
|
"grad_norm": 8.602317810058594, |
|
"learning_rate": 1.74e-06, |
|
"loss": 0.7629, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.33507853403141363, |
|
"grad_norm": 2.7594990730285645, |
|
"learning_rate": 1.9900000000000004e-06, |
|
"loss": 0.7418, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3769633507853403, |
|
"grad_norm": 2.7047348022460938, |
|
"learning_rate": 2.24e-06, |
|
"loss": 0.7166, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.418848167539267, |
|
"grad_norm": 2.0666158199310303, |
|
"learning_rate": 2.4900000000000003e-06, |
|
"loss": 0.6876, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.418848167539267, |
|
"eval_loss": 0.5937426090240479, |
|
"eval_runtime": 263.0955, |
|
"eval_samples_per_second": 32.266, |
|
"eval_steps_per_second": 4.037, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4607329842931937, |
|
"grad_norm": 2.955193519592285, |
|
"learning_rate": 2.7400000000000004e-06, |
|
"loss": 0.673, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5026178010471204, |
|
"grad_norm": 4.856322765350342, |
|
"learning_rate": 2.99e-06, |
|
"loss": 0.6539, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5445026178010471, |
|
"grad_norm": 7.631103992462158, |
|
"learning_rate": 3.2400000000000003e-06, |
|
"loss": 0.6588, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5863874345549738, |
|
"grad_norm": 2.715771436691284, |
|
"learning_rate": 3.49e-06, |
|
"loss": 0.624, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6282722513089005, |
|
"grad_norm": 2.7684273719787598, |
|
"learning_rate": 3.74e-06, |
|
"loss": 0.6116, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6701570680628273, |
|
"grad_norm": 2.219599723815918, |
|
"learning_rate": 3.990000000000001e-06, |
|
"loss": 0.6097, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7120418848167539, |
|
"grad_norm": 3.749188184738159, |
|
"learning_rate": 4.24e-06, |
|
"loss": 0.614, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7539267015706806, |
|
"grad_norm": 2.120619773864746, |
|
"learning_rate": 4.49e-06, |
|
"loss": 0.5814, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7958115183246073, |
|
"grad_norm": 2.7725327014923096, |
|
"learning_rate": 4.74e-06, |
|
"loss": 0.5895, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.837696335078534, |
|
"grad_norm": 9.851961135864258, |
|
"learning_rate": 4.9900000000000005e-06, |
|
"loss": 0.5623, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.837696335078534, |
|
"eval_loss": 0.4933945834636688, |
|
"eval_runtime": 258.739, |
|
"eval_samples_per_second": 32.809, |
|
"eval_steps_per_second": 4.105, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8795811518324608, |
|
"grad_norm": 5.074089527130127, |
|
"learning_rate": 5.240000000000001e-06, |
|
"loss": 0.5748, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.9214659685863874, |
|
"grad_norm": 5.884639739990234, |
|
"learning_rate": 5.485e-06, |
|
"loss": 0.5395, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9633507853403142, |
|
"grad_norm": 3.1588447093963623, |
|
"learning_rate": 5.735e-06, |
|
"loss": 0.5478, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.0052356020942408, |
|
"grad_norm": 2.4452970027923584, |
|
"learning_rate": 5.985000000000001e-06, |
|
"loss": 0.538, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.0471204188481675, |
|
"grad_norm": 8.290769577026367, |
|
"learning_rate": 6.235000000000001e-06, |
|
"loss": 0.5411, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.0890052356020943, |
|
"grad_norm": 4.080046653747559, |
|
"learning_rate": 6.485000000000001e-06, |
|
"loss": 0.5387, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.130890052356021, |
|
"grad_norm": 2.7520201206207275, |
|
"learning_rate": 6.735000000000001e-06, |
|
"loss": 0.5426, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.1727748691099475, |
|
"grad_norm": 3.8192107677459717, |
|
"learning_rate": 6.985000000000001e-06, |
|
"loss": 0.5315, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.2146596858638743, |
|
"grad_norm": 9.535676956176758, |
|
"learning_rate": 7.235000000000001e-06, |
|
"loss": 0.5159, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.256544502617801, |
|
"grad_norm": 2.5731022357940674, |
|
"learning_rate": 7.485000000000001e-06, |
|
"loss": 0.5165, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.256544502617801, |
|
"eval_loss": 0.4628298878669739, |
|
"eval_runtime": 257.0265, |
|
"eval_samples_per_second": 33.028, |
|
"eval_steps_per_second": 4.132, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.2984293193717278, |
|
"grad_norm": 3.662090539932251, |
|
"learning_rate": 7.735e-06, |
|
"loss": 0.5188, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.3403141361256545, |
|
"grad_norm": 3.1680572032928467, |
|
"learning_rate": 7.985e-06, |
|
"loss": 0.5148, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.3821989528795813, |
|
"grad_norm": 3.601471185684204, |
|
"learning_rate": 8.235e-06, |
|
"loss": 0.5127, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.4240837696335078, |
|
"grad_norm": 4.497046947479248, |
|
"learning_rate": 8.485000000000001e-06, |
|
"loss": 0.5039, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.4659685863874345, |
|
"grad_norm": 3.6496424674987793, |
|
"learning_rate": 8.735000000000002e-06, |
|
"loss": 0.5114, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.5078534031413613, |
|
"grad_norm": 2.7080061435699463, |
|
"learning_rate": 8.985000000000001e-06, |
|
"loss": 0.4953, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.5497382198952878, |
|
"grad_norm": 2.5439274311065674, |
|
"learning_rate": 9.235e-06, |
|
"loss": 0.497, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.5916230366492146, |
|
"grad_norm": 2.004573106765747, |
|
"learning_rate": 9.485000000000002e-06, |
|
"loss": 0.4889, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.6335078534031413, |
|
"grad_norm": 4.699184894561768, |
|
"learning_rate": 9.735e-06, |
|
"loss": 0.4925, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.675392670157068, |
|
"grad_norm": 3.2080323696136475, |
|
"learning_rate": 9.985000000000002e-06, |
|
"loss": 0.4851, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.675392670157068, |
|
"eval_loss": 0.43240800499916077, |
|
"eval_runtime": 256.4745, |
|
"eval_samples_per_second": 33.099, |
|
"eval_steps_per_second": 4.141, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.7172774869109948, |
|
"grad_norm": 2.5463037490844727, |
|
"learning_rate": 9.953000000000001e-06, |
|
"loss": 0.4802, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.7591623036649215, |
|
"grad_norm": 3.6697189807891846, |
|
"learning_rate": 9.903e-06, |
|
"loss": 0.485, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.8010471204188483, |
|
"grad_norm": 2.6455466747283936, |
|
"learning_rate": 9.853e-06, |
|
"loss": 0.4863, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.8429319371727748, |
|
"grad_norm": 2.849780321121216, |
|
"learning_rate": 9.803e-06, |
|
"loss": 0.4824, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.8848167539267016, |
|
"grad_norm": 2.037757158279419, |
|
"learning_rate": 9.753e-06, |
|
"loss": 0.4762, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.9267015706806283, |
|
"grad_norm": 2.060755491256714, |
|
"learning_rate": 9.703000000000002e-06, |
|
"loss": 0.4793, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.9685863874345548, |
|
"grad_norm": 2.3721907138824463, |
|
"learning_rate": 9.653e-06, |
|
"loss": 0.4744, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.0104712041884816, |
|
"grad_norm": 2.605724573135376, |
|
"learning_rate": 9.603000000000001e-06, |
|
"loss": 0.4665, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.0523560209424083, |
|
"grad_norm": 1.9419960975646973, |
|
"learning_rate": 9.553000000000002e-06, |
|
"loss": 0.4701, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.094240837696335, |
|
"grad_norm": 2.3074657917022705, |
|
"learning_rate": 9.503e-06, |
|
"loss": 0.4672, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.094240837696335, |
|
"eval_loss": 0.4174647927284241, |
|
"eval_runtime": 259.1813, |
|
"eval_samples_per_second": 32.753, |
|
"eval_steps_per_second": 4.098, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.136125654450262, |
|
"grad_norm": 20.03424644470215, |
|
"learning_rate": 9.453e-06, |
|
"loss": 0.4676, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.1780104712041886, |
|
"grad_norm": 3.410041093826294, |
|
"learning_rate": 9.403000000000001e-06, |
|
"loss": 0.471, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.2198952879581153, |
|
"grad_norm": 2.1835362911224365, |
|
"learning_rate": 9.353000000000002e-06, |
|
"loss": 0.4639, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.261780104712042, |
|
"grad_norm": 2.245004177093506, |
|
"learning_rate": 9.303e-06, |
|
"loss": 0.4642, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.303664921465969, |
|
"grad_norm": 2.2452590465545654, |
|
"learning_rate": 9.253000000000001e-06, |
|
"loss": 0.4638, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.345549738219895, |
|
"grad_norm": 3.5843396186828613, |
|
"learning_rate": 9.203000000000002e-06, |
|
"loss": 0.4546, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.387434554973822, |
|
"grad_norm": 1.689903974533081, |
|
"learning_rate": 9.153e-06, |
|
"loss": 0.4607, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.4293193717277486, |
|
"grad_norm": 2.26529860496521, |
|
"learning_rate": 9.103e-06, |
|
"loss": 0.4732, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.4712041884816753, |
|
"grad_norm": 1.669374942779541, |
|
"learning_rate": 9.053000000000001e-06, |
|
"loss": 0.4578, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.513089005235602, |
|
"grad_norm": 2.6614444255828857, |
|
"learning_rate": 9.003e-06, |
|
"loss": 0.4615, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.513089005235602, |
|
"eval_loss": 0.40531307458877563, |
|
"eval_runtime": 260.7294, |
|
"eval_samples_per_second": 32.559, |
|
"eval_steps_per_second": 4.073, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.554973821989529, |
|
"grad_norm": 2.6586270332336426, |
|
"learning_rate": 8.953e-06, |
|
"loss": 0.4486, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.5968586387434556, |
|
"grad_norm": 1.5337822437286377, |
|
"learning_rate": 8.903000000000001e-06, |
|
"loss": 0.4474, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.6387434554973823, |
|
"grad_norm": 5.568012714385986, |
|
"learning_rate": 8.853e-06, |
|
"loss": 0.4466, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.680628272251309, |
|
"grad_norm": 1.8297160863876343, |
|
"learning_rate": 8.803e-06, |
|
"loss": 0.454, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.7225130890052354, |
|
"grad_norm": 3.0298142433166504, |
|
"learning_rate": 8.753e-06, |
|
"loss": 0.4513, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.7643979057591626, |
|
"grad_norm": 2.6482794284820557, |
|
"learning_rate": 8.703e-06, |
|
"loss": 0.4461, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.806282722513089, |
|
"grad_norm": 1.9304325580596924, |
|
"learning_rate": 8.653e-06, |
|
"loss": 0.4464, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.8481675392670156, |
|
"grad_norm": 2.3229286670684814, |
|
"learning_rate": 8.603e-06, |
|
"loss": 0.4475, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.8900523560209423, |
|
"grad_norm": 1.8658236265182495, |
|
"learning_rate": 8.553000000000001e-06, |
|
"loss": 0.4476, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.931937172774869, |
|
"grad_norm": 1.574404239654541, |
|
"learning_rate": 8.503e-06, |
|
"loss": 0.4426, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.931937172774869, |
|
"eval_loss": 0.39671415090560913, |
|
"eval_runtime": 259.1033, |
|
"eval_samples_per_second": 32.763, |
|
"eval_steps_per_second": 4.099, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.973821989528796, |
|
"grad_norm": 2.0523784160614014, |
|
"learning_rate": 8.453000000000002e-06, |
|
"loss": 0.4545, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 3.0157068062827226, |
|
"grad_norm": 2.5431597232818604, |
|
"learning_rate": 8.403e-06, |
|
"loss": 0.4416, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.0575916230366493, |
|
"grad_norm": 1.7480727434158325, |
|
"learning_rate": 8.353000000000001e-06, |
|
"loss": 0.4396, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 3.099476439790576, |
|
"grad_norm": 2.19262433052063, |
|
"learning_rate": 8.303000000000002e-06, |
|
"loss": 0.4403, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.141361256544503, |
|
"grad_norm": 1.6949037313461304, |
|
"learning_rate": 8.253e-06, |
|
"loss": 0.4345, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 3.183246073298429, |
|
"grad_norm": 3.506704807281494, |
|
"learning_rate": 8.203000000000001e-06, |
|
"loss": 0.4364, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.225130890052356, |
|
"grad_norm": 2.155595064163208, |
|
"learning_rate": 8.153000000000001e-06, |
|
"loss": 0.4361, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 3.2670157068062826, |
|
"grad_norm": 1.7618342638015747, |
|
"learning_rate": 8.103e-06, |
|
"loss": 0.4348, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.3089005235602094, |
|
"grad_norm": 2.2834861278533936, |
|
"learning_rate": 8.053e-06, |
|
"loss": 0.4435, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 3.350785340314136, |
|
"grad_norm": 1.7081154584884644, |
|
"learning_rate": 8.003000000000001e-06, |
|
"loss": 0.4295, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.350785340314136, |
|
"eval_loss": 0.389825701713562, |
|
"eval_runtime": 254.7511, |
|
"eval_samples_per_second": 33.323, |
|
"eval_steps_per_second": 4.169, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.392670157068063, |
|
"grad_norm": 2.6197593212127686, |
|
"learning_rate": 7.953e-06, |
|
"loss": 0.4312, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 3.4345549738219896, |
|
"grad_norm": 1.742619514465332, |
|
"learning_rate": 7.903e-06, |
|
"loss": 0.432, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.4764397905759163, |
|
"grad_norm": 1.730943202972412, |
|
"learning_rate": 7.853000000000001e-06, |
|
"loss": 0.4376, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 3.518324607329843, |
|
"grad_norm": 2.4986751079559326, |
|
"learning_rate": 7.803000000000001e-06, |
|
"loss": 0.4344, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.5602094240837694, |
|
"grad_norm": 1.707566499710083, |
|
"learning_rate": 7.753e-06, |
|
"loss": 0.435, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 3.6020942408376966, |
|
"grad_norm": 2.230375051498413, |
|
"learning_rate": 7.703e-06, |
|
"loss": 0.4286, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.643979057591623, |
|
"grad_norm": 2.0870070457458496, |
|
"learning_rate": 7.653000000000001e-06, |
|
"loss": 0.4302, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 3.6858638743455496, |
|
"grad_norm": 2.5888140201568604, |
|
"learning_rate": 7.603000000000001e-06, |
|
"loss": 0.4333, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.7277486910994764, |
|
"grad_norm": 3.0323987007141113, |
|
"learning_rate": 7.553e-06, |
|
"loss": 0.4296, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 3.769633507853403, |
|
"grad_norm": 2.430225133895874, |
|
"learning_rate": 7.503e-06, |
|
"loss": 0.4323, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.769633507853403, |
|
"eval_loss": 0.38729745149612427, |
|
"eval_runtime": 254.688, |
|
"eval_samples_per_second": 33.331, |
|
"eval_steps_per_second": 4.17, |
|
"step": 4500 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 12000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 11, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.04828814811703e+16, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|