|
{ |
|
"best_metric": 0.6268656716417911, |
|
"best_model_checkpoint": "videomae-base-finetuned-good-gesturePhaseV1/checkpoint-276", |
|
"epoch": 3.239130434782609, |
|
"eval_steps": 500, |
|
"global_step": 276, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.036231884057971016, |
|
"grad_norm": 3.9660768508911133, |
|
"learning_rate": 1.785714285714286e-05, |
|
"loss": 1.1407, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.07246376811594203, |
|
"grad_norm": 4.645390033721924, |
|
"learning_rate": 3.571428571428572e-05, |
|
"loss": 1.1465, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10869565217391304, |
|
"grad_norm": 4.135717868804932, |
|
"learning_rate": 4.959677419354839e-05, |
|
"loss": 1.1081, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.14492753623188406, |
|
"grad_norm": 5.431738376617432, |
|
"learning_rate": 4.7580645161290326e-05, |
|
"loss": 1.0511, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18115942028985507, |
|
"grad_norm": 7.023768901824951, |
|
"learning_rate": 4.556451612903226e-05, |
|
"loss": 1.262, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.21739130434782608, |
|
"grad_norm": 2.841621160507202, |
|
"learning_rate": 4.3548387096774194e-05, |
|
"loss": 0.9855, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.2536231884057971, |
|
"grad_norm": 5.4796671867370605, |
|
"learning_rate": 4.1532258064516135e-05, |
|
"loss": 1.2297, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.2536231884057971, |
|
"eval_accuracy": 0.5597014925373134, |
|
"eval_accuracy_hold": 0.0, |
|
"eval_accuracy_preparation": 1.0, |
|
"eval_accuracy_recovery": 0.0, |
|
"eval_accuracy_stroke": 0.0, |
|
"eval_accuracy_unknown": 0.0, |
|
"eval_loss": 1.2679108381271362, |
|
"eval_runtime": 11.7358, |
|
"eval_samples_per_second": 11.418, |
|
"eval_steps_per_second": 1.449, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.036231884057971, |
|
"grad_norm": 6.009732246398926, |
|
"learning_rate": 3.951612903225806e-05, |
|
"loss": 0.9358, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.0724637681159421, |
|
"grad_norm": 5.398038864135742, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 1.3549, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.108695652173913, |
|
"grad_norm": 3.8027758598327637, |
|
"learning_rate": 3.548387096774194e-05, |
|
"loss": 1.1816, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.144927536231884, |
|
"grad_norm": 5.348720550537109, |
|
"learning_rate": 3.346774193548387e-05, |
|
"loss": 1.1078, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.181159420289855, |
|
"grad_norm": 4.720065593719482, |
|
"learning_rate": 3.1451612903225806e-05, |
|
"loss": 1.2226, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.2173913043478262, |
|
"grad_norm": 5.930412292480469, |
|
"learning_rate": 2.9435483870967743e-05, |
|
"loss": 1.1558, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.2536231884057971, |
|
"grad_norm": 5.240575790405273, |
|
"learning_rate": 2.7419354838709678e-05, |
|
"loss": 0.9724, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.2536231884057971, |
|
"eval_accuracy": 0.5597014925373134, |
|
"eval_accuracy_hold": 0.0, |
|
"eval_accuracy_preparation": 1.0, |
|
"eval_accuracy_recovery": 0.0, |
|
"eval_accuracy_stroke": 0.0, |
|
"eval_accuracy_unknown": 0.0, |
|
"eval_loss": 1.2204008102416992, |
|
"eval_runtime": 11.5396, |
|
"eval_samples_per_second": 11.612, |
|
"eval_steps_per_second": 1.473, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.036231884057971, |
|
"grad_norm": 3.1463561058044434, |
|
"learning_rate": 2.5403225806451615e-05, |
|
"loss": 1.035, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.072463768115942, |
|
"grad_norm": 7.061142444610596, |
|
"learning_rate": 2.338709677419355e-05, |
|
"loss": 1.1356, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.108695652173913, |
|
"grad_norm": 3.777250289916992, |
|
"learning_rate": 2.1370967741935487e-05, |
|
"loss": 0.8868, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.1449275362318843, |
|
"grad_norm": 13.91281509399414, |
|
"learning_rate": 1.935483870967742e-05, |
|
"loss": 1.1425, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.181159420289855, |
|
"grad_norm": 6.732293128967285, |
|
"learning_rate": 1.733870967741936e-05, |
|
"loss": 1.1183, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.217391304347826, |
|
"grad_norm": 7.715019226074219, |
|
"learning_rate": 1.5322580645161292e-05, |
|
"loss": 1.1121, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.253623188405797, |
|
"grad_norm": 7.341396331787109, |
|
"learning_rate": 1.3306451612903225e-05, |
|
"loss": 0.9151, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.253623188405797, |
|
"eval_accuracy": 0.6119402985074627, |
|
"eval_accuracy_hold": 0.0, |
|
"eval_accuracy_preparation": 0.9333333333333333, |
|
"eval_accuracy_recovery": 0.0, |
|
"eval_accuracy_stroke": 0.0, |
|
"eval_accuracy_unknown": 0.46153846153846156, |
|
"eval_loss": 1.0775725841522217, |
|
"eval_runtime": 15.5605, |
|
"eval_samples_per_second": 8.612, |
|
"eval_steps_per_second": 1.093, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.036231884057971, |
|
"grad_norm": 4.763265132904053, |
|
"learning_rate": 1.129032258064516e-05, |
|
"loss": 1.1173, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.072463768115942, |
|
"grad_norm": 12.301189422607422, |
|
"learning_rate": 9.274193548387097e-06, |
|
"loss": 0.8348, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.108695652173913, |
|
"grad_norm": 5.9293036460876465, |
|
"learning_rate": 7.258064516129033e-06, |
|
"loss": 0.8975, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.1449275362318843, |
|
"grad_norm": 7.083910942077637, |
|
"learning_rate": 5.241935483870968e-06, |
|
"loss": 0.9537, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.181159420289855, |
|
"grad_norm": 7.554308891296387, |
|
"learning_rate": 3.225806451612903e-06, |
|
"loss": 1.0105, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.217391304347826, |
|
"grad_norm": 9.840630531311035, |
|
"learning_rate": 1.2096774193548388e-06, |
|
"loss": 0.9169, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.239130434782609, |
|
"eval_accuracy": 0.6268656716417911, |
|
"eval_accuracy_hold": 0.0, |
|
"eval_accuracy_preparation": 0.9066666666666666, |
|
"eval_accuracy_recovery": 0.0, |
|
"eval_accuracy_stroke": 0.0, |
|
"eval_accuracy_unknown": 0.6153846153846154, |
|
"eval_loss": 1.0385252237319946, |
|
"eval_runtime": 13.3185, |
|
"eval_samples_per_second": 10.061, |
|
"eval_steps_per_second": 1.276, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 3.239130434782609, |
|
"step": 276, |
|
"total_flos": 2.74390763571413e+18, |
|
"train_loss": 1.068202899849933, |
|
"train_runtime": 497.3862, |
|
"train_samples_per_second": 4.439, |
|
"train_steps_per_second": 0.555 |
|
}, |
|
{ |
|
"epoch": 3.239130434782609, |
|
"eval_accuracy": 0.5918367346938775, |
|
"eval_accuracy_hold": 0.0, |
|
"eval_accuracy_preparation": 0.922077922077922, |
|
"eval_accuracy_recovery": 0.0, |
|
"eval_accuracy_stroke": 0.0, |
|
"eval_accuracy_unknown": 0.45714285714285713, |
|
"eval_loss": 1.1627883911132812, |
|
"eval_runtime": 14.7864, |
|
"eval_samples_per_second": 9.942, |
|
"eval_steps_per_second": 1.285, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 3.239130434782609, |
|
"eval_accuracy": 0.5918367346938775, |
|
"eval_accuracy_hold": 0.0, |
|
"eval_accuracy_preparation": 0.922077922077922, |
|
"eval_accuracy_recovery": 0.0, |
|
"eval_accuracy_stroke": 0.0, |
|
"eval_accuracy_unknown": 0.45714285714285713, |
|
"eval_loss": 1.1627883911132812, |
|
"eval_runtime": 13.4243, |
|
"eval_samples_per_second": 10.95, |
|
"eval_steps_per_second": 1.415, |
|
"step": 276 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 276, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 9223372036854775807, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.74390763571413e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|