|
{ |
|
"best_metric": 0.289421480911704, |
|
"best_model_checkpoint": "/cluster/home/torstefl/Master/saved_model/W2V/Combined/BB-NB-22.05/checkpoint-17670", |
|
"epoch": 40.0, |
|
"eval_steps": 500, |
|
"global_step": 18600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 5.088616847991943, |
|
"learning_rate": 3.086666666666667e-05, |
|
"loss": 2.0999, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_rundkast_loss": 0.39486202597618103, |
|
"eval_rundkast_runtime": 32.0081, |
|
"eval_rundkast_samples_per_second": 42.021, |
|
"eval_rundkast_steps_per_second": 1.343, |
|
"eval_rundkast_wer": 0.22597836461979, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_NB_BB_loss": 1.3563096523284912, |
|
"eval_NB_BB_runtime": 84.5302, |
|
"eval_NB_BB_samples_per_second": 22.513, |
|
"eval_NB_BB_steps_per_second": 0.71, |
|
"eval_NB_BB_wer": 0.40965952795433386, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_nb_samtale_loss": 0.5479292273521423, |
|
"eval_nb_samtale_runtime": 36.6651, |
|
"eval_nb_samtale_samples_per_second": 14.537, |
|
"eval_nb_samtale_steps_per_second": 0.464, |
|
"eval_nb_samtale_wer": 0.2421875, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_bigbrother_loss": 1.670620083808899, |
|
"eval_bigbrother_runtime": 39.7215, |
|
"eval_bigbrother_samples_per_second": 34.49, |
|
"eval_bigbrother_steps_per_second": 1.083, |
|
"eval_bigbrother_wer": 0.5755256585837428, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 5.0017313957214355, |
|
"learning_rate": 6.186666666666668e-05, |
|
"loss": 1.7526, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_rundkast_loss": 0.3949491083621979, |
|
"eval_rundkast_runtime": 32.7201, |
|
"eval_rundkast_samples_per_second": 41.106, |
|
"eval_rundkast_steps_per_second": 1.314, |
|
"eval_rundkast_wer": 0.22398981864460707, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_NB_BB_loss": 1.2307980060577393, |
|
"eval_NB_BB_runtime": 83.9537, |
|
"eval_NB_BB_samples_per_second": 22.667, |
|
"eval_NB_BB_steps_per_second": 0.715, |
|
"eval_NB_BB_wer": 0.38634063398242985, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_nb_samtale_loss": 0.5440447330474854, |
|
"eval_nb_samtale_runtime": 36.2064, |
|
"eval_nb_samtale_samples_per_second": 14.721, |
|
"eval_nb_samtale_steps_per_second": 0.47, |
|
"eval_nb_samtale_wer": 0.23177083333333334, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_bigbrother_loss": 1.4979978799819946, |
|
"eval_bigbrother_runtime": 39.598, |
|
"eval_bigbrother_samples_per_second": 34.598, |
|
"eval_bigbrother_steps_per_second": 1.086, |
|
"eval_bigbrother_wer": 0.5396761459759929, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 7.238979816436768, |
|
"learning_rate": 9.286666666666667e-05, |
|
"loss": 1.6611, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_rundkast_loss": 0.4249923825263977, |
|
"eval_rundkast_runtime": 32.4044, |
|
"eval_rundkast_samples_per_second": 41.507, |
|
"eval_rundkast_steps_per_second": 1.327, |
|
"eval_rundkast_wer": 0.22621699013681196, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_NB_BB_loss": 1.2449074983596802, |
|
"eval_NB_BB_runtime": 85.2007, |
|
"eval_NB_BB_samples_per_second": 22.335, |
|
"eval_NB_BB_steps_per_second": 0.704, |
|
"eval_NB_BB_wer": 0.3861786972187361, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_nb_samtale_loss": 0.5582835078239441, |
|
"eval_nb_samtale_runtime": 36.2595, |
|
"eval_nb_samtale_samples_per_second": 14.7, |
|
"eval_nb_samtale_steps_per_second": 0.469, |
|
"eval_nb_samtale_wer": 0.23274739583333334, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_bigbrother_loss": 1.5122398138046265, |
|
"eval_bigbrother_runtime": 39.5302, |
|
"eval_bigbrother_samples_per_second": 34.657, |
|
"eval_bigbrother_steps_per_second": 1.088, |
|
"eval_bigbrother_wer": 0.5397567066784822, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 9.255756378173828, |
|
"learning_rate": 9.790643274853801e-05, |
|
"loss": 1.5765, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_rundkast_loss": 0.37671104073524475, |
|
"eval_rundkast_runtime": 31.5626, |
|
"eval_rundkast_samples_per_second": 42.614, |
|
"eval_rundkast_steps_per_second": 1.362, |
|
"eval_rundkast_wer": 0.220967228762329, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_NB_BB_loss": 1.0437960624694824, |
|
"eval_NB_BB_runtime": 84.5888, |
|
"eval_NB_BB_samples_per_second": 22.497, |
|
"eval_NB_BB_steps_per_second": 0.709, |
|
"eval_NB_BB_wer": 0.3624954455285211, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_nb_samtale_loss": 0.48440223932266235, |
|
"eval_nb_samtale_runtime": 39.3769, |
|
"eval_nb_samtale_samples_per_second": 13.536, |
|
"eval_nb_samtale_steps_per_second": 0.432, |
|
"eval_nb_samtale_wer": 0.22314453125, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_bigbrother_loss": 1.2615869045257568, |
|
"eval_bigbrother_runtime": 41.8866, |
|
"eval_bigbrother_samples_per_second": 32.707, |
|
"eval_bigbrother_steps_per_second": 1.027, |
|
"eval_bigbrother_wer": 0.5008458873761379, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 6.270685195922852, |
|
"learning_rate": 9.519883040935674e-05, |
|
"loss": 1.4853, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_rundkast_loss": 0.3830512762069702, |
|
"eval_rundkast_runtime": 31.7506, |
|
"eval_rundkast_samples_per_second": 42.361, |
|
"eval_rundkast_steps_per_second": 1.354, |
|
"eval_rundkast_wer": 0.22351256761056315, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_NB_BB_loss": 1.0370458364486694, |
|
"eval_NB_BB_runtime": 83.374, |
|
"eval_NB_BB_samples_per_second": 22.825, |
|
"eval_NB_BB_steps_per_second": 0.72, |
|
"eval_NB_BB_wer": 0.35549168049876523, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_nb_samtale_loss": 0.47176724672317505, |
|
"eval_nb_samtale_runtime": 35.9414, |
|
"eval_nb_samtale_samples_per_second": 14.83, |
|
"eval_nb_samtale_steps_per_second": 0.473, |
|
"eval_nb_samtale_wer": 0.21818033854166666, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_bigbrother_loss": 1.256990671157837, |
|
"eval_bigbrother_runtime": 39.4041, |
|
"eval_bigbrother_samples_per_second": 34.768, |
|
"eval_bigbrother_steps_per_second": 1.091, |
|
"eval_bigbrother_wer": 0.49093692096995084, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 6.830248832702637, |
|
"learning_rate": 9.24795321637427e-05, |
|
"loss": 1.4316, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_rundkast_loss": 0.41122955083847046, |
|
"eval_rundkast_runtime": 31.5822, |
|
"eval_rundkast_samples_per_second": 42.587, |
|
"eval_rundkast_steps_per_second": 1.362, |
|
"eval_rundkast_wer": 0.2255011135857461, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_NB_BB_loss": 1.0182600021362305, |
|
"eval_NB_BB_runtime": 83.3578, |
|
"eval_NB_BB_samples_per_second": 22.829, |
|
"eval_NB_BB_steps_per_second": 0.72, |
|
"eval_NB_BB_wer": 0.3526577871341241, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_nb_samtale_loss": 0.4850824475288391, |
|
"eval_nb_samtale_runtime": 35.9277, |
|
"eval_nb_samtale_samples_per_second": 14.835, |
|
"eval_nb_samtale_steps_per_second": 0.473, |
|
"eval_nb_samtale_wer": 0.22054036458333334, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_bigbrother_loss": 1.2255350351333618, |
|
"eval_bigbrother_runtime": 39.6894, |
|
"eval_bigbrother_samples_per_second": 34.518, |
|
"eval_bigbrother_steps_per_second": 1.083, |
|
"eval_bigbrother_wer": 0.4833642149359542, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 5.503568172454834, |
|
"learning_rate": 8.976023391812865e-05, |
|
"loss": 1.3699, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_rundkast_loss": 0.409690260887146, |
|
"eval_rundkast_runtime": 31.88, |
|
"eval_rundkast_samples_per_second": 42.189, |
|
"eval_rundkast_steps_per_second": 1.349, |
|
"eval_rundkast_wer": 0.2182628062360802, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_NB_BB_loss": 0.9981555342674255, |
|
"eval_NB_BB_runtime": 83.1726, |
|
"eval_NB_BB_samples_per_second": 22.88, |
|
"eval_NB_BB_steps_per_second": 0.721, |
|
"eval_NB_BB_wer": 0.33933848832031094, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_nb_samtale_loss": 0.4907973110675812, |
|
"eval_nb_samtale_runtime": 35.9575, |
|
"eval_nb_samtale_samples_per_second": 14.823, |
|
"eval_nb_samtale_steps_per_second": 0.473, |
|
"eval_nb_samtale_wer": 0.21346028645833334, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_bigbrother_loss": 1.1955358982086182, |
|
"eval_bigbrother_runtime": 39.6796, |
|
"eval_bigbrother_samples_per_second": 34.527, |
|
"eval_bigbrother_steps_per_second": 1.084, |
|
"eval_bigbrother_wer": 0.46435188914847336, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 11.339502334594727, |
|
"learning_rate": 8.704093567251463e-05, |
|
"loss": 1.3346, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_rundkast_loss": 0.3831312358379364, |
|
"eval_rundkast_runtime": 32.9308, |
|
"eval_rundkast_samples_per_second": 40.843, |
|
"eval_rundkast_steps_per_second": 1.306, |
|
"eval_rundkast_wer": 0.2166719694559338, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_NB_BB_loss": 0.9318952560424805, |
|
"eval_NB_BB_runtime": 98.1107, |
|
"eval_NB_BB_samples_per_second": 19.396, |
|
"eval_NB_BB_steps_per_second": 0.612, |
|
"eval_NB_BB_wer": 0.3345613537913445, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_nb_samtale_loss": 0.458311527967453, |
|
"eval_nb_samtale_runtime": 36.9177, |
|
"eval_nb_samtale_samples_per_second": 14.438, |
|
"eval_nb_samtale_steps_per_second": 0.46, |
|
"eval_nb_samtale_wer": 0.20939127604166666, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_bigbrother_loss": 1.1163750886917114, |
|
"eval_bigbrother_runtime": 40.3476, |
|
"eval_bigbrother_samples_per_second": 33.955, |
|
"eval_bigbrother_steps_per_second": 1.066, |
|
"eval_bigbrother_wer": 0.4583903971642633, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 2.8908886909484863, |
|
"learning_rate": 8.432163742690059e-05, |
|
"loss": 1.2933, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_rundkast_loss": 0.3958103656768799, |
|
"eval_rundkast_runtime": 32.4676, |
|
"eval_rundkast_samples_per_second": 41.426, |
|
"eval_rundkast_steps_per_second": 1.324, |
|
"eval_rundkast_wer": 0.21563792554883868, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_NB_BB_loss": 0.9900856614112854, |
|
"eval_NB_BB_runtime": 85.2037, |
|
"eval_NB_BB_samples_per_second": 22.335, |
|
"eval_NB_BB_steps_per_second": 0.704, |
|
"eval_NB_BB_wer": 0.33201084976316747, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_nb_samtale_loss": 0.48341503739356995, |
|
"eval_nb_samtale_runtime": 36.9893, |
|
"eval_nb_samtale_samples_per_second": 14.41, |
|
"eval_nb_samtale_steps_per_second": 0.46, |
|
"eval_nb_samtale_wer": 0.206298828125, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_bigbrother_loss": 1.1871527433395386, |
|
"eval_bigbrother_runtime": 41.4962, |
|
"eval_bigbrother_samples_per_second": 33.015, |
|
"eval_bigbrother_steps_per_second": 1.036, |
|
"eval_bigbrother_wer": 0.4568597438169661, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 5.720297336578369, |
|
"learning_rate": 8.160233918128655e-05, |
|
"loss": 1.2483, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_rundkast_loss": 0.4080227017402649, |
|
"eval_rundkast_runtime": 32.4095, |
|
"eval_rundkast_samples_per_second": 41.5, |
|
"eval_rundkast_steps_per_second": 1.327, |
|
"eval_rundkast_wer": 0.21181991727648744, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_NB_BB_loss": 0.9500575065612793, |
|
"eval_NB_BB_runtime": 84.6361, |
|
"eval_NB_BB_samples_per_second": 22.485, |
|
"eval_NB_BB_steps_per_second": 0.709, |
|
"eval_NB_BB_wer": 0.3251285373061819, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_nb_samtale_loss": 0.4818229675292969, |
|
"eval_nb_samtale_runtime": 37.1509, |
|
"eval_nb_samtale_samples_per_second": 14.347, |
|
"eval_nb_samtale_steps_per_second": 0.458, |
|
"eval_nb_samtale_wer": 0.201904296875, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_bigbrother_loss": 1.1322485208511353, |
|
"eval_bigbrother_runtime": 40.3892, |
|
"eval_bigbrother_samples_per_second": 33.92, |
|
"eval_bigbrother_steps_per_second": 1.065, |
|
"eval_bigbrother_wer": 0.44638685249335375, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"grad_norm": 15.0436372756958, |
|
"learning_rate": 7.888304093567252e-05, |
|
"loss": 1.2131, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_rundkast_loss": 0.43607091903686523, |
|
"eval_rundkast_runtime": 32.6212, |
|
"eval_rundkast_samples_per_second": 41.231, |
|
"eval_rundkast_steps_per_second": 1.318, |
|
"eval_rundkast_wer": 0.21492204899777284, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_NB_BB_loss": 1.0018919706344604, |
|
"eval_NB_BB_runtime": 85.111, |
|
"eval_NB_BB_samples_per_second": 22.359, |
|
"eval_NB_BB_steps_per_second": 0.705, |
|
"eval_NB_BB_wer": 0.32488563216064126, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_nb_samtale_loss": 0.4951498210430145, |
|
"eval_nb_samtale_runtime": 36.7127, |
|
"eval_nb_samtale_samples_per_second": 14.518, |
|
"eval_nb_samtale_steps_per_second": 0.463, |
|
"eval_nb_samtale_wer": 0.2001953125, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_bigbrother_loss": 1.1989638805389404, |
|
"eval_bigbrother_runtime": 40.7234, |
|
"eval_bigbrother_samples_per_second": 33.642, |
|
"eval_bigbrother_steps_per_second": 1.056, |
|
"eval_bigbrother_wer": 0.4492064770804801, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 6.106450080871582, |
|
"learning_rate": 7.616374269005848e-05, |
|
"loss": 1.1821, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_rundkast_loss": 0.41126009821891785, |
|
"eval_rundkast_runtime": 33.0331, |
|
"eval_rundkast_samples_per_second": 40.717, |
|
"eval_rundkast_steps_per_second": 1.302, |
|
"eval_rundkast_wer": 0.2180241807190582, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_NB_BB_loss": 0.9629638195037842, |
|
"eval_NB_BB_runtime": 99.548, |
|
"eval_NB_BB_samples_per_second": 19.116, |
|
"eval_NB_BB_steps_per_second": 0.603, |
|
"eval_NB_BB_wer": 0.3250880531152585, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_nb_samtale_loss": 0.46501702070236206, |
|
"eval_nb_samtale_runtime": 36.9075, |
|
"eval_nb_samtale_samples_per_second": 14.442, |
|
"eval_nb_samtale_steps_per_second": 0.461, |
|
"eval_nb_samtale_wer": 0.202880859375, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_bigbrother_loss": 1.1566274166107178, |
|
"eval_bigbrother_runtime": 40.4961, |
|
"eval_bigbrother_samples_per_second": 33.83, |
|
"eval_bigbrother_steps_per_second": 1.062, |
|
"eval_bigbrother_wer": 0.4453395633609925, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"grad_norm": 10.281853675842285, |
|
"learning_rate": 7.344444444444445e-05, |
|
"loss": 1.1652, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_rundkast_loss": 0.38290590047836304, |
|
"eval_rundkast_runtime": 32.4453, |
|
"eval_rundkast_samples_per_second": 41.454, |
|
"eval_rundkast_steps_per_second": 1.325, |
|
"eval_rundkast_wer": 0.21563792554883868, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_NB_BB_loss": 0.9116414785385132, |
|
"eval_NB_BB_runtime": 84.9456, |
|
"eval_NB_BB_samples_per_second": 22.403, |
|
"eval_NB_BB_steps_per_second": 0.706, |
|
"eval_NB_BB_wer": 0.31994656086798107, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_nb_samtale_loss": 0.4675864279270172, |
|
"eval_nb_samtale_runtime": 36.7336, |
|
"eval_nb_samtale_samples_per_second": 14.51, |
|
"eval_nb_samtale_steps_per_second": 0.463, |
|
"eval_nb_samtale_wer": 0.198486328125, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_bigbrother_loss": 1.0843714475631714, |
|
"eval_bigbrother_runtime": 40.2808, |
|
"eval_bigbrother_samples_per_second": 34.011, |
|
"eval_bigbrother_steps_per_second": 1.068, |
|
"eval_bigbrother_wer": 0.4403447998066543, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"grad_norm": 3.9103004932403564, |
|
"learning_rate": 7.072514619883041e-05, |
|
"loss": 1.1179, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_rundkast_loss": 0.3869563937187195, |
|
"eval_rundkast_runtime": 32.4154, |
|
"eval_rundkast_samples_per_second": 41.493, |
|
"eval_rundkast_steps_per_second": 1.327, |
|
"eval_rundkast_wer": 0.2121380846325167, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_NB_BB_loss": 0.9029076099395752, |
|
"eval_NB_BB_runtime": 85.8744, |
|
"eval_NB_BB_samples_per_second": 22.16, |
|
"eval_NB_BB_steps_per_second": 0.699, |
|
"eval_NB_BB_wer": 0.31063519695558883, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_nb_samtale_loss": 0.4522353410720825, |
|
"eval_nb_samtale_runtime": 36.6303, |
|
"eval_nb_samtale_samples_per_second": 14.551, |
|
"eval_nb_samtale_steps_per_second": 0.464, |
|
"eval_nb_samtale_wer": 0.19287109375, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_bigbrother_loss": 1.0779362916946411, |
|
"eval_bigbrother_runtime": 40.3876, |
|
"eval_bigbrother_samples_per_second": 33.921, |
|
"eval_bigbrother_steps_per_second": 1.065, |
|
"eval_bigbrother_wer": 0.42818013373076613, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"grad_norm": 6.248087406158447, |
|
"learning_rate": 6.800584795321637e-05, |
|
"loss": 1.1123, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_rundkast_loss": 0.43526676297187805, |
|
"eval_rundkast_runtime": 32.5572, |
|
"eval_rundkast_samples_per_second": 41.312, |
|
"eval_rundkast_steps_per_second": 1.321, |
|
"eval_rundkast_wer": 0.21110404072542158, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_NB_BB_loss": 1.0294309854507446, |
|
"eval_NB_BB_runtime": 85.0774, |
|
"eval_NB_BB_samples_per_second": 22.368, |
|
"eval_NB_BB_steps_per_second": 0.705, |
|
"eval_NB_BB_wer": 0.3128618274563783, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_nb_samtale_loss": 0.5030134916305542, |
|
"eval_nb_samtale_runtime": 36.6885, |
|
"eval_nb_samtale_samples_per_second": 14.528, |
|
"eval_nb_samtale_steps_per_second": 0.463, |
|
"eval_nb_samtale_wer": 0.193603515625, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_bigbrother_loss": 1.2345293760299683, |
|
"eval_bigbrother_runtime": 40.3974, |
|
"eval_bigbrother_samples_per_second": 33.913, |
|
"eval_bigbrother_steps_per_second": 1.064, |
|
"eval_bigbrother_wer": 0.4302747119954886, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 6.180894374847412, |
|
"learning_rate": 6.529239766081871e-05, |
|
"loss": 1.075, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_rundkast_loss": 0.4082247316837311, |
|
"eval_rundkast_runtime": 32.5615, |
|
"eval_rundkast_samples_per_second": 41.306, |
|
"eval_rundkast_steps_per_second": 1.321, |
|
"eval_rundkast_wer": 0.21006999681832644, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_NB_BB_loss": 0.9210716485977173, |
|
"eval_NB_BB_runtime": 84.9343, |
|
"eval_NB_BB_samples_per_second": 22.406, |
|
"eval_NB_BB_steps_per_second": 0.706, |
|
"eval_NB_BB_wer": 0.3099064815189668, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_nb_samtale_loss": 0.459839791059494, |
|
"eval_nb_samtale_runtime": 37.1878, |
|
"eval_nb_samtale_samples_per_second": 14.333, |
|
"eval_nb_samtale_steps_per_second": 0.457, |
|
"eval_nb_samtale_wer": 0.19303385416666666, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_bigbrother_loss": 1.100581407546997, |
|
"eval_bigbrother_runtime": 40.4416, |
|
"eval_bigbrother_samples_per_second": 33.876, |
|
"eval_bigbrother_steps_per_second": 1.063, |
|
"eval_bigbrother_wer": 0.42487714492870376, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"grad_norm": 5.789996147155762, |
|
"learning_rate": 6.257309941520468e-05, |
|
"loss": 1.0718, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_rundkast_loss": 0.4331410825252533, |
|
"eval_rundkast_runtime": 32.5516, |
|
"eval_rundkast_samples_per_second": 41.319, |
|
"eval_rundkast_steps_per_second": 1.321, |
|
"eval_rundkast_wer": 0.21014953865733377, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_NB_BB_loss": 0.965010404586792, |
|
"eval_NB_BB_runtime": 84.6139, |
|
"eval_NB_BB_samples_per_second": 22.49, |
|
"eval_NB_BB_steps_per_second": 0.709, |
|
"eval_NB_BB_wer": 0.3072345249180195, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_nb_samtale_loss": 0.49670931696891785, |
|
"eval_nb_samtale_runtime": 36.4484, |
|
"eval_nb_samtale_samples_per_second": 14.623, |
|
"eval_nb_samtale_steps_per_second": 0.466, |
|
"eval_nb_samtale_wer": 0.19140625, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_bigbrother_loss": 1.1466903686523438, |
|
"eval_bigbrother_runtime": 40.4761, |
|
"eval_bigbrother_samples_per_second": 33.847, |
|
"eval_bigbrother_steps_per_second": 1.062, |
|
"eval_bigbrother_wer": 0.4217352775316201, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"grad_norm": 12.003623008728027, |
|
"learning_rate": 5.985380116959065e-05, |
|
"loss": 1.0422, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_rundkast_loss": 0.4282049536705017, |
|
"eval_rundkast_runtime": 32.7905, |
|
"eval_rundkast_samples_per_second": 41.018, |
|
"eval_rundkast_steps_per_second": 1.311, |
|
"eval_rundkast_wer": 0.21134266624244352, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_NB_BB_loss": 0.9594568610191345, |
|
"eval_NB_BB_runtime": 84.0609, |
|
"eval_NB_BB_samples_per_second": 22.638, |
|
"eval_NB_BB_steps_per_second": 0.714, |
|
"eval_NB_BB_wer": 0.30707258815432575, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_nb_samtale_loss": 0.4719415307044983, |
|
"eval_nb_samtale_runtime": 36.4276, |
|
"eval_nb_samtale_samples_per_second": 14.632, |
|
"eval_nb_samtale_steps_per_second": 0.467, |
|
"eval_nb_samtale_wer": 0.192626953125, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_bigbrother_loss": 1.148333191871643, |
|
"eval_bigbrother_runtime": 40.3155, |
|
"eval_bigbrother_samples_per_second": 33.982, |
|
"eval_bigbrother_steps_per_second": 1.067, |
|
"eval_bigbrother_wer": 0.42068798839925886, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"grad_norm": 7.242124080657959, |
|
"learning_rate": 5.7134502923976606e-05, |
|
"loss": 1.0303, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_rundkast_loss": 0.41116222739219666, |
|
"eval_rundkast_runtime": 31.7737, |
|
"eval_rundkast_samples_per_second": 42.331, |
|
"eval_rundkast_steps_per_second": 1.353, |
|
"eval_rundkast_wer": 0.21293350302258987, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_NB_BB_loss": 0.9378307461738586, |
|
"eval_NB_BB_runtime": 94.3567, |
|
"eval_NB_BB_samples_per_second": 20.168, |
|
"eval_NB_BB_steps_per_second": 0.636, |
|
"eval_NB_BB_wer": 0.3072345249180195, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_nb_samtale_loss": 0.46919578313827515, |
|
"eval_nb_samtale_runtime": 35.5991, |
|
"eval_nb_samtale_samples_per_second": 14.972, |
|
"eval_nb_samtale_steps_per_second": 0.478, |
|
"eval_nb_samtale_wer": 0.19327799479166666, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_bigbrother_loss": 1.1198594570159912, |
|
"eval_bigbrother_runtime": 39.6114, |
|
"eval_bigbrother_samples_per_second": 34.586, |
|
"eval_bigbrother_steps_per_second": 1.086, |
|
"eval_bigbrother_wer": 0.4197212599693869, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 6.254478931427002, |
|
"learning_rate": 5.4415204678362576e-05, |
|
"loss": 1.0192, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_rundkast_loss": 0.4086886942386627, |
|
"eval_rundkast_runtime": 32.5071, |
|
"eval_rundkast_samples_per_second": 41.376, |
|
"eval_rundkast_steps_per_second": 1.323, |
|
"eval_rundkast_wer": 0.20752465797009226, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_NB_BB_loss": 0.9522827863693237, |
|
"eval_NB_BB_runtime": 100.9277, |
|
"eval_NB_BB_samples_per_second": 18.855, |
|
"eval_NB_BB_steps_per_second": 0.594, |
|
"eval_NB_BB_wer": 0.3014048014250435, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_nb_samtale_loss": 0.4663674533367157, |
|
"eval_nb_samtale_runtime": 35.8162, |
|
"eval_nb_samtale_samples_per_second": 14.882, |
|
"eval_nb_samtale_steps_per_second": 0.475, |
|
"eval_nb_samtale_wer": 0.18782552083333334, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_bigbrother_loss": 1.1411128044128418, |
|
"eval_bigbrother_runtime": 39.8861, |
|
"eval_bigbrother_samples_per_second": 34.348, |
|
"eval_bigbrother_steps_per_second": 1.078, |
|
"eval_bigbrother_wer": 0.4136792072826875, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"grad_norm": 11.991388320922852, |
|
"learning_rate": 5.169590643274854e-05, |
|
"loss": 1.0087, |
|
"step": 9765 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_rundkast_loss": 0.41283562779426575, |
|
"eval_rundkast_runtime": 31.8401, |
|
"eval_rundkast_samples_per_second": 42.242, |
|
"eval_rundkast_steps_per_second": 1.351, |
|
"eval_rundkast_wer": 0.20919503658924593, |
|
"step": 9765 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_NB_BB_loss": 0.9513369202613831, |
|
"eval_NB_BB_runtime": 83.7994, |
|
"eval_NB_BB_samples_per_second": 22.709, |
|
"eval_NB_BB_steps_per_second": 0.716, |
|
"eval_NB_BB_wer": 0.30274077972551716, |
|
"step": 9765 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_nb_samtale_loss": 0.47523027658462524, |
|
"eval_nb_samtale_runtime": 35.727, |
|
"eval_nb_samtale_samples_per_second": 14.919, |
|
"eval_nb_samtale_steps_per_second": 0.476, |
|
"eval_nb_samtale_wer": 0.18977864583333334, |
|
"step": 9765 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_bigbrother_loss": 1.1363465785980225, |
|
"eval_bigbrother_runtime": 39.6372, |
|
"eval_bigbrother_samples_per_second": 34.564, |
|
"eval_bigbrother_steps_per_second": 1.085, |
|
"eval_bigbrother_wer": 0.4141625714976235, |
|
"step": 9765 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"grad_norm": 8.134933471679688, |
|
"learning_rate": 4.8976608187134504e-05, |
|
"loss": 0.982, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_rundkast_loss": 0.4441734850406647, |
|
"eval_rundkast_runtime": 31.8611, |
|
"eval_rundkast_samples_per_second": 42.214, |
|
"eval_rundkast_steps_per_second": 1.35, |
|
"eval_rundkast_wer": 0.20696786509704104, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_NB_BB_loss": 1.0050716400146484, |
|
"eval_NB_BB_runtime": 83.4701, |
|
"eval_NB_BB_samples_per_second": 22.799, |
|
"eval_NB_BB_steps_per_second": 0.719, |
|
"eval_NB_BB_wer": 0.30059511760657465, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_nb_samtale_loss": 0.4712933897972107, |
|
"eval_nb_samtale_runtime": 35.6009, |
|
"eval_nb_samtale_samples_per_second": 14.972, |
|
"eval_nb_samtale_steps_per_second": 0.478, |
|
"eval_nb_samtale_wer": 0.18782552083333334, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_bigbrother_loss": 1.213335394859314, |
|
"eval_bigbrother_runtime": 39.7468, |
|
"eval_bigbrother_samples_per_second": 34.468, |
|
"eval_bigbrother_steps_per_second": 1.082, |
|
"eval_bigbrother_wer": 0.41247079674534765, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"grad_norm": 33.663204193115234, |
|
"learning_rate": 4.625730994152047e-05, |
|
"loss": 0.9667, |
|
"step": 10695 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_rundkast_loss": 0.43352487683296204, |
|
"eval_rundkast_runtime": 32.6485, |
|
"eval_rundkast_samples_per_second": 41.196, |
|
"eval_rundkast_steps_per_second": 1.317, |
|
"eval_rundkast_wer": 0.2063315303849825, |
|
"step": 10695 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_NB_BB_loss": 0.9717670679092407, |
|
"eval_NB_BB_runtime": 84.3242, |
|
"eval_NB_BB_samples_per_second": 22.568, |
|
"eval_NB_BB_steps_per_second": 0.712, |
|
"eval_NB_BB_wer": 0.2978826768147039, |
|
"step": 10695 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_nb_samtale_loss": 0.451678067445755, |
|
"eval_nb_samtale_runtime": 36.8422, |
|
"eval_nb_samtale_samples_per_second": 14.467, |
|
"eval_nb_samtale_steps_per_second": 0.461, |
|
"eval_nb_samtale_wer": 0.18570963541666666, |
|
"step": 10695 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_bigbrother_loss": 1.173987865447998, |
|
"eval_bigbrother_runtime": 40.426, |
|
"eval_bigbrother_samples_per_second": 33.889, |
|
"eval_bigbrother_steps_per_second": 1.064, |
|
"eval_bigbrother_wer": 0.408362200918392, |
|
"step": 10695 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"grad_norm": 3.15181827545166, |
|
"learning_rate": 4.354385964912281e-05, |
|
"loss": 0.9586, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_rundkast_loss": 0.43574872612953186, |
|
"eval_rundkast_runtime": 34.0532, |
|
"eval_rundkast_samples_per_second": 39.497, |
|
"eval_rundkast_steps_per_second": 1.263, |
|
"eval_rundkast_wer": 0.20776328348711423, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_NB_BB_loss": 0.9633657932281494, |
|
"eval_NB_BB_runtime": 85.5761, |
|
"eval_NB_BB_samples_per_second": 22.238, |
|
"eval_NB_BB_steps_per_second": 0.701, |
|
"eval_NB_BB_wer": 0.2978421926237804, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_nb_samtale_loss": 0.45962560176849365, |
|
"eval_nb_samtale_runtime": 36.3588, |
|
"eval_nb_samtale_samples_per_second": 14.659, |
|
"eval_nb_samtale_steps_per_second": 0.468, |
|
"eval_nb_samtale_wer": 0.186767578125, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_bigbrother_loss": 1.159108281135559, |
|
"eval_bigbrother_runtime": 40.6416, |
|
"eval_bigbrother_samples_per_second": 33.709, |
|
"eval_bigbrother_steps_per_second": 1.058, |
|
"eval_bigbrother_wer": 0.4082816402159027, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"grad_norm": 15.837915420532227, |
|
"learning_rate": 4.082456140350878e-05, |
|
"loss": 0.9366, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_rundkast_loss": 0.43792787194252014, |
|
"eval_rundkast_runtime": 32.8971, |
|
"eval_rundkast_samples_per_second": 40.885, |
|
"eval_rundkast_steps_per_second": 1.307, |
|
"eval_rundkast_wer": 0.20664969774101177, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_NB_BB_loss": 1.0051846504211426, |
|
"eval_NB_BB_runtime": 85.0128, |
|
"eval_NB_BB_samples_per_second": 22.385, |
|
"eval_NB_BB_steps_per_second": 0.706, |
|
"eval_NB_BB_wer": 0.2959394356503785, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_nb_samtale_loss": 0.478113055229187, |
|
"eval_nb_samtale_runtime": 36.2152, |
|
"eval_nb_samtale_samples_per_second": 14.718, |
|
"eval_nb_samtale_steps_per_second": 0.469, |
|
"eval_nb_samtale_wer": 0.18636067708333334, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_bigbrother_loss": 1.2104246616363525, |
|
"eval_bigbrother_runtime": 40.5472, |
|
"eval_bigbrother_samples_per_second": 33.788, |
|
"eval_bigbrother_steps_per_second": 1.06, |
|
"eval_bigbrother_wer": 0.404656408603883, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"grad_norm": 3.201197624206543, |
|
"learning_rate": 3.8105263157894735e-05, |
|
"loss": 0.9325, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_rundkast_loss": 0.41804516315460205, |
|
"eval_rundkast_runtime": 31.7914, |
|
"eval_rundkast_samples_per_second": 42.307, |
|
"eval_rundkast_steps_per_second": 1.353, |
|
"eval_rundkast_wer": 0.2049793191218581, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_NB_BB_loss": 0.9606844186782837, |
|
"eval_NB_BB_runtime": 83.5304, |
|
"eval_NB_BB_samples_per_second": 22.782, |
|
"eval_NB_BB_steps_per_second": 0.718, |
|
"eval_NB_BB_wer": 0.2959394356503785, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_nb_samtale_loss": 0.44722747802734375, |
|
"eval_nb_samtale_runtime": 35.6931, |
|
"eval_nb_samtale_samples_per_second": 14.933, |
|
"eval_nb_samtale_steps_per_second": 0.476, |
|
"eval_nb_samtale_wer": 0.18611653645833334, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_bigbrother_loss": 1.160502314567566, |
|
"eval_bigbrother_runtime": 39.8504, |
|
"eval_bigbrother_samples_per_second": 34.379, |
|
"eval_bigbrother_steps_per_second": 1.079, |
|
"eval_bigbrother_wer": 0.4047369693063724, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"grad_norm": 5.304738521575928, |
|
"learning_rate": 3.5385964912280706e-05, |
|
"loss": 0.9189, |
|
"step": 12555 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_rundkast_loss": 0.4250892698764801, |
|
"eval_rundkast_runtime": 31.7351, |
|
"eval_rundkast_samples_per_second": 42.382, |
|
"eval_rundkast_steps_per_second": 1.355, |
|
"eval_rundkast_wer": 0.20641107222398983, |
|
"step": 12555 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_NB_BB_loss": 0.9551168084144592, |
|
"eval_NB_BB_runtime": 83.3377, |
|
"eval_NB_BB_samples_per_second": 22.835, |
|
"eval_NB_BB_steps_per_second": 0.72, |
|
"eval_NB_BB_wer": 0.29614185660499576, |
|
"step": 12555 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_nb_samtale_loss": 0.4538469910621643, |
|
"eval_nb_samtale_runtime": 35.5954, |
|
"eval_nb_samtale_samples_per_second": 14.974, |
|
"eval_nb_samtale_steps_per_second": 0.478, |
|
"eval_nb_samtale_wer": 0.18416341145833334, |
|
"step": 12555 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_bigbrother_loss": 1.149815559387207, |
|
"eval_bigbrother_runtime": 39.7731, |
|
"eval_bigbrother_samples_per_second": 34.445, |
|
"eval_bigbrother_steps_per_second": 1.081, |
|
"eval_bigbrother_wer": 0.40634818335615885, |
|
"step": 12555 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"grad_norm": 20.373756408691406, |
|
"learning_rate": 3.266666666666667e-05, |
|
"loss": 0.9033, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_rundkast_loss": 0.42664840817451477, |
|
"eval_rundkast_runtime": 31.8414, |
|
"eval_rundkast_samples_per_second": 42.241, |
|
"eval_rundkast_steps_per_second": 1.35, |
|
"eval_rundkast_wer": 0.2053770283168947, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_NB_BB_loss": 0.9819029569625854, |
|
"eval_NB_BB_runtime": 83.3501, |
|
"eval_NB_BB_samples_per_second": 22.831, |
|
"eval_NB_BB_steps_per_second": 0.72, |
|
"eval_NB_BB_wer": 0.2954536253592972, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_nb_samtale_loss": 0.4682862460613251, |
|
"eval_nb_samtale_runtime": 35.6354, |
|
"eval_nb_samtale_samples_per_second": 14.957, |
|
"eval_nb_samtale_steps_per_second": 0.477, |
|
"eval_nb_samtale_wer": 0.185791015625, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_bigbrother_loss": 1.1819084882736206, |
|
"eval_bigbrother_runtime": 39.7653, |
|
"eval_bigbrother_samples_per_second": 34.452, |
|
"eval_bigbrother_steps_per_second": 1.081, |
|
"eval_bigbrother_wer": 0.4032868766615645, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"grad_norm": 2.7378759384155273, |
|
"learning_rate": 2.9947368421052634e-05, |
|
"loss": 0.9018, |
|
"step": 13485 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_rundkast_loss": 0.4359777867794037, |
|
"eval_rundkast_runtime": 31.7585, |
|
"eval_rundkast_samples_per_second": 42.351, |
|
"eval_rundkast_steps_per_second": 1.354, |
|
"eval_rundkast_wer": 0.20553611199490932, |
|
"step": 13485 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_NB_BB_loss": 0.9609730243682861, |
|
"eval_NB_BB_runtime": 83.7711, |
|
"eval_NB_BB_samples_per_second": 22.717, |
|
"eval_NB_BB_steps_per_second": 0.716, |
|
"eval_NB_BB_wer": 0.2929840897129671, |
|
"step": 13485 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_nb_samtale_loss": 0.4618481695652008, |
|
"eval_nb_samtale_runtime": 36.0874, |
|
"eval_nb_samtale_samples_per_second": 14.77, |
|
"eval_nb_samtale_steps_per_second": 0.471, |
|
"eval_nb_samtale_wer": 0.18489583333333334, |
|
"step": 13485 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_bigbrother_loss": 1.1553022861480713, |
|
"eval_bigbrother_runtime": 39.8384, |
|
"eval_bigbrother_samples_per_second": 34.389, |
|
"eval_bigbrother_steps_per_second": 1.079, |
|
"eval_bigbrother_wer": 0.3999838878595021, |
|
"step": 13485 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"grad_norm": 27.257183074951172, |
|
"learning_rate": 2.7228070175438598e-05, |
|
"loss": 0.8937, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_rundkast_loss": 0.45607230067253113, |
|
"eval_rundkast_runtime": 32.367, |
|
"eval_rundkast_samples_per_second": 41.555, |
|
"eval_rundkast_steps_per_second": 1.329, |
|
"eval_rundkast_wer": 0.20688832325803372, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_NB_BB_loss": 1.0051723718643188, |
|
"eval_NB_BB_runtime": 103.4936, |
|
"eval_NB_BB_samples_per_second": 18.388, |
|
"eval_NB_BB_steps_per_second": 0.58, |
|
"eval_NB_BB_wer": 0.294522488968058, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_nb_samtale_loss": 0.470134973526001, |
|
"eval_nb_samtale_runtime": 36.186, |
|
"eval_nb_samtale_samples_per_second": 14.729, |
|
"eval_nb_samtale_steps_per_second": 0.47, |
|
"eval_nb_samtale_wer": 0.18440755208333334, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_bigbrother_loss": 1.2132840156555176, |
|
"eval_bigbrother_runtime": 40.683, |
|
"eval_bigbrother_samples_per_second": 33.675, |
|
"eval_bigbrother_steps_per_second": 1.057, |
|
"eval_bigbrother_wer": 0.4032063159590752, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"grad_norm": 5.251443862915039, |
|
"learning_rate": 2.4508771929824562e-05, |
|
"loss": 0.8814, |
|
"step": 14415 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_rundkast_loss": 0.4411740005016327, |
|
"eval_rundkast_runtime": 32.5448, |
|
"eval_rundkast_samples_per_second": 41.328, |
|
"eval_rundkast_steps_per_second": 1.321, |
|
"eval_rundkast_wer": 0.20617244670696785, |
|
"step": 14415 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_NB_BB_loss": 0.9743751287460327, |
|
"eval_NB_BB_runtime": 84.3758, |
|
"eval_NB_BB_samples_per_second": 22.554, |
|
"eval_NB_BB_steps_per_second": 0.711, |
|
"eval_NB_BB_wer": 0.2928626371401968, |
|
"step": 14415 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_nb_samtale_loss": 0.47333696484565735, |
|
"eval_nb_samtale_runtime": 36.2272, |
|
"eval_nb_samtale_samples_per_second": 14.713, |
|
"eval_nb_samtale_steps_per_second": 0.469, |
|
"eval_nb_samtale_wer": 0.18359375, |
|
"step": 14415 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_bigbrother_loss": 1.1694875955581665, |
|
"eval_bigbrother_runtime": 40.7287, |
|
"eval_bigbrother_samples_per_second": 33.637, |
|
"eval_bigbrother_steps_per_second": 1.056, |
|
"eval_bigbrother_wer": 0.4003061306694594, |
|
"step": 14415 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"grad_norm": 5.162887096405029, |
|
"learning_rate": 2.1789473684210526e-05, |
|
"loss": 0.8784, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_rundkast_loss": 0.45319777727127075, |
|
"eval_rundkast_runtime": 33.2279, |
|
"eval_rundkast_samples_per_second": 40.478, |
|
"eval_rundkast_steps_per_second": 1.294, |
|
"eval_rundkast_wer": 0.2041043588927776, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_NB_BB_loss": 1.0061391592025757, |
|
"eval_NB_BB_runtime": 85.8491, |
|
"eval_NB_BB_samples_per_second": 22.167, |
|
"eval_NB_BB_steps_per_second": 0.699, |
|
"eval_NB_BB_wer": 0.2929031213311202, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_nb_samtale_loss": 0.486108660697937, |
|
"eval_nb_samtale_runtime": 35.9271, |
|
"eval_nb_samtale_samples_per_second": 14.836, |
|
"eval_nb_samtale_steps_per_second": 0.473, |
|
"eval_nb_samtale_wer": 0.18400065104166666, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_bigbrother_loss": 1.208513855934143, |
|
"eval_bigbrother_runtime": 40.5432, |
|
"eval_bigbrother_samples_per_second": 33.791, |
|
"eval_bigbrother_steps_per_second": 1.061, |
|
"eval_bigbrother_wer": 0.4007894948843954, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"grad_norm": 5.579722881317139, |
|
"learning_rate": 1.907017543859649e-05, |
|
"loss": 0.8831, |
|
"step": 15345 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_rundkast_loss": 0.44335484504699707, |
|
"eval_rundkast_runtime": 32.683, |
|
"eval_rundkast_samples_per_second": 41.153, |
|
"eval_rundkast_steps_per_second": 1.316, |
|
"eval_rundkast_wer": 0.2053770283168947, |
|
"step": 15345 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_NB_BB_loss": 1.0044219493865967, |
|
"eval_NB_BB_runtime": 85.3142, |
|
"eval_NB_BB_samples_per_second": 22.306, |
|
"eval_NB_BB_steps_per_second": 0.703, |
|
"eval_NB_BB_wer": 0.2922958584672685, |
|
"step": 15345 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_nb_samtale_loss": 0.4627678096294403, |
|
"eval_nb_samtale_runtime": 36.1313, |
|
"eval_nb_samtale_samples_per_second": 14.752, |
|
"eval_nb_samtale_steps_per_second": 0.471, |
|
"eval_nb_samtale_wer": 0.18367513020833334, |
|
"step": 15345 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_bigbrother_loss": 1.2153956890106201, |
|
"eval_bigbrother_runtime": 40.81, |
|
"eval_bigbrother_samples_per_second": 33.57, |
|
"eval_bigbrother_steps_per_second": 1.054, |
|
"eval_bigbrother_wer": 0.40070893418190606, |
|
"step": 15345 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"grad_norm": 5.071865081787109, |
|
"learning_rate": 1.6356725146198832e-05, |
|
"loss": 0.8722, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_rundkast_loss": 0.4348403215408325, |
|
"eval_rundkast_runtime": 32.9227, |
|
"eval_rundkast_samples_per_second": 40.853, |
|
"eval_rundkast_steps_per_second": 1.306, |
|
"eval_rundkast_wer": 0.20521794463888005, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_NB_BB_loss": 0.9682305455207825, |
|
"eval_NB_BB_runtime": 85.5917, |
|
"eval_NB_BB_samples_per_second": 22.233, |
|
"eval_NB_BB_steps_per_second": 0.701, |
|
"eval_NB_BB_wer": 0.2904335856847901, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_nb_samtale_loss": 0.44841158390045166, |
|
"eval_nb_samtale_runtime": 36.1939, |
|
"eval_nb_samtale_samples_per_second": 14.726, |
|
"eval_nb_samtale_steps_per_second": 0.47, |
|
"eval_nb_samtale_wer": 0.18212890625, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_bigbrother_loss": 1.170255422592163, |
|
"eval_bigbrother_runtime": 40.7341, |
|
"eval_bigbrother_samples_per_second": 33.633, |
|
"eval_bigbrother_steps_per_second": 1.056, |
|
"eval_bigbrother_wer": 0.39813099170224764, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"grad_norm": 8.128209114074707, |
|
"learning_rate": 1.3637426900584796e-05, |
|
"loss": 0.859, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_rundkast_loss": 0.446013867855072, |
|
"eval_rundkast_runtime": 32.6724, |
|
"eval_rundkast_samples_per_second": 41.166, |
|
"eval_rundkast_steps_per_second": 1.316, |
|
"eval_rundkast_wer": 0.20243398027362391, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_NB_BB_loss": 1.003732442855835, |
|
"eval_NB_BB_runtime": 84.4708, |
|
"eval_NB_BB_samples_per_second": 22.528, |
|
"eval_NB_BB_steps_per_second": 0.71, |
|
"eval_NB_BB_wer": 0.29023116473017285, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_nb_samtale_loss": 0.4694906175136566, |
|
"eval_nb_samtale_runtime": 35.8944, |
|
"eval_nb_samtale_samples_per_second": 14.849, |
|
"eval_nb_samtale_steps_per_second": 0.474, |
|
"eval_nb_samtale_wer": 0.18359375, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_bigbrother_loss": 1.2112008333206177, |
|
"eval_bigbrother_runtime": 40.7858, |
|
"eval_bigbrother_samples_per_second": 33.59, |
|
"eval_bigbrother_steps_per_second": 1.054, |
|
"eval_bigbrother_wer": 0.39635865624748245, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"grad_norm": 7.049020767211914, |
|
"learning_rate": 1.091812865497076e-05, |
|
"loss": 0.8508, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_rundkast_loss": 0.4368053376674652, |
|
"eval_rundkast_runtime": 32.6393, |
|
"eval_rundkast_samples_per_second": 41.208, |
|
"eval_rundkast_steps_per_second": 1.317, |
|
"eval_rundkast_wer": 0.20330894050270443, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_NB_BB_loss": 0.9765028357505798, |
|
"eval_NB_BB_runtime": 86.1031, |
|
"eval_NB_BB_samples_per_second": 22.101, |
|
"eval_NB_BB_steps_per_second": 0.697, |
|
"eval_NB_BB_wer": 0.29031213311201975, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_nb_samtale_loss": 0.45226001739501953, |
|
"eval_nb_samtale_runtime": 36.3397, |
|
"eval_nb_samtale_samples_per_second": 14.667, |
|
"eval_nb_samtale_steps_per_second": 0.468, |
|
"eval_nb_samtale_wer": 0.182373046875, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_bigbrother_loss": 1.180801272392273, |
|
"eval_bigbrother_runtime": 41.007, |
|
"eval_bigbrother_samples_per_second": 33.409, |
|
"eval_bigbrother_steps_per_second": 1.049, |
|
"eval_bigbrother_wer": 0.3970031418673971, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"grad_norm": 8.417683601379395, |
|
"learning_rate": 8.198830409356726e-06, |
|
"loss": 0.8547, |
|
"step": 17205 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_rundkast_loss": 0.44972118735313416, |
|
"eval_rundkast_runtime": 32.9953, |
|
"eval_rundkast_samples_per_second": 40.763, |
|
"eval_rundkast_steps_per_second": 1.303, |
|
"eval_rundkast_wer": 0.20291123130766783, |
|
"step": 17205 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_NB_BB_loss": 1.0050333738327026, |
|
"eval_NB_BB_runtime": 85.63, |
|
"eval_NB_BB_samples_per_second": 22.224, |
|
"eval_NB_BB_steps_per_second": 0.701, |
|
"eval_NB_BB_wer": 0.29002874377555565, |
|
"step": 17205 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_nb_samtale_loss": 0.4681548774242401, |
|
"eval_nb_samtale_runtime": 35.9123, |
|
"eval_nb_samtale_samples_per_second": 14.842, |
|
"eval_nb_samtale_steps_per_second": 0.473, |
|
"eval_nb_samtale_wer": 0.18212890625, |
|
"step": 17205 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_bigbrother_loss": 1.2139179706573486, |
|
"eval_bigbrother_runtime": 40.7816, |
|
"eval_bigbrother_samples_per_second": 33.594, |
|
"eval_bigbrother_steps_per_second": 1.054, |
|
"eval_bigbrother_wer": 0.3976476274873117, |
|
"step": 17205 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"grad_norm": 6.699246883392334, |
|
"learning_rate": 5.479532163742691e-06, |
|
"loss": 0.8443, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_rundkast_loss": 0.4439980089664459, |
|
"eval_rundkast_runtime": 32.4126, |
|
"eval_rundkast_samples_per_second": 41.496, |
|
"eval_rundkast_steps_per_second": 1.327, |
|
"eval_rundkast_wer": 0.20370664969774102, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_NB_BB_loss": 0.9929541349411011, |
|
"eval_NB_BB_runtime": 85.6285, |
|
"eval_NB_BB_samples_per_second": 22.224, |
|
"eval_NB_BB_steps_per_second": 0.701, |
|
"eval_NB_BB_wer": 0.289421480911704, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_nb_samtale_loss": 0.46245497465133667, |
|
"eval_nb_samtale_runtime": 36.3757, |
|
"eval_nb_samtale_samples_per_second": 14.653, |
|
"eval_nb_samtale_steps_per_second": 0.467, |
|
"eval_nb_samtale_wer": 0.18229166666666666, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_bigbrother_loss": 1.1993677616119385, |
|
"eval_bigbrother_runtime": 40.9524, |
|
"eval_bigbrother_samples_per_second": 33.453, |
|
"eval_bigbrother_steps_per_second": 1.05, |
|
"eval_bigbrother_wer": 0.3961975348425038, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"grad_norm": 5.469863414764404, |
|
"learning_rate": 2.760233918128655e-06, |
|
"loss": 0.835, |
|
"step": 18135 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_rundkast_loss": 0.44508007168769836, |
|
"eval_rundkast_runtime": 32.7027, |
|
"eval_rundkast_samples_per_second": 41.128, |
|
"eval_rundkast_steps_per_second": 1.315, |
|
"eval_rundkast_wer": 0.2041043588927776, |
|
"step": 18135 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_NB_BB_loss": 0.9969344139099121, |
|
"eval_NB_BB_runtime": 84.86, |
|
"eval_NB_BB_samples_per_second": 22.425, |
|
"eval_NB_BB_steps_per_second": 0.707, |
|
"eval_NB_BB_wer": 0.29023116473017285, |
|
"step": 18135 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_nb_samtale_loss": 0.4631071090698242, |
|
"eval_nb_samtale_runtime": 35.8723, |
|
"eval_nb_samtale_samples_per_second": 14.858, |
|
"eval_nb_samtale_steps_per_second": 0.474, |
|
"eval_nb_samtale_wer": 0.18180338541666666, |
|
"step": 18135 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_bigbrother_loss": 1.2046737670898438, |
|
"eval_bigbrother_runtime": 40.6261, |
|
"eval_bigbrother_samples_per_second": 33.722, |
|
"eval_bigbrother_steps_per_second": 1.058, |
|
"eval_bigbrother_wer": 0.3973253846773544, |
|
"step": 18135 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 2.851428508758545, |
|
"learning_rate": 4.678362573099416e-08, |
|
"loss": 0.8346, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_rundkast_loss": 0.44796615839004517, |
|
"eval_rundkast_runtime": 32.5088, |
|
"eval_rundkast_samples_per_second": 41.373, |
|
"eval_rundkast_steps_per_second": 1.323, |
|
"eval_rundkast_wer": 0.2036271078587337, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_NB_BB_loss": 1.0030395984649658, |
|
"eval_NB_BB_runtime": 94.7646, |
|
"eval_NB_BB_samples_per_second": 20.081, |
|
"eval_NB_BB_steps_per_second": 0.633, |
|
"eval_NB_BB_wer": 0.28958341767539775, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_nb_samtale_loss": 0.4656994044780731, |
|
"eval_nb_samtale_runtime": 36.739, |
|
"eval_nb_samtale_samples_per_second": 14.508, |
|
"eval_nb_samtale_steps_per_second": 0.463, |
|
"eval_nb_samtale_wer": 0.18123372395833334, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_bigbrother_loss": 1.2117459774017334, |
|
"eval_bigbrother_runtime": 40.801, |
|
"eval_bigbrother_samples_per_second": 33.578, |
|
"eval_bigbrother_steps_per_second": 1.054, |
|
"eval_bigbrother_wer": 0.3970031418673971, |
|
"step": 18600 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 18600, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 40, |
|
"save_steps": 500, |
|
"total_flos": 1.281422114476163e+20, |
|
"train_batch_size": 48, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|