|
{ |
|
"best_metric": 0.6592887043952942, |
|
"best_model_checkpoint": "saves/Gemma-7B/lora/train_1/checkpoint-110", |
|
"epoch": 0.17500683620453925, |
|
"eval_steps": 10, |
|
"global_step": 160, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.010937927262783703, |
|
"grad_norm": 16.269868850708008, |
|
"learning_rate": 0.00029999015487222375, |
|
"loss": 3.1586, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.010937927262783703, |
|
"eval_loss": 0.7969412207603455, |
|
"eval_runtime": 1611.4779, |
|
"eval_samples_per_second": 29.24, |
|
"eval_steps_per_second": 0.914, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.021875854525567406, |
|
"grad_norm": 1.8665393590927124, |
|
"learning_rate": 0.00029996062078124905, |
|
"loss": 0.6717, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.021875854525567406, |
|
"eval_loss": 0.7616342306137085, |
|
"eval_runtime": 1610.0951, |
|
"eval_samples_per_second": 29.265, |
|
"eval_steps_per_second": 0.915, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03281378178835111, |
|
"grad_norm": 2.046764373779297, |
|
"learning_rate": 0.0002999114016039678, |
|
"loss": 0.5741, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03281378178835111, |
|
"eval_loss": 0.7397875189781189, |
|
"eval_runtime": 1609.4859, |
|
"eval_samples_per_second": 29.276, |
|
"eval_steps_per_second": 0.915, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.04375170905113481, |
|
"grad_norm": 0.8763797879219055, |
|
"learning_rate": 0.00029984250380130117, |
|
"loss": 0.5809, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04375170905113481, |
|
"eval_loss": 0.7744876742362976, |
|
"eval_runtime": 1610.4604, |
|
"eval_samples_per_second": 29.259, |
|
"eval_steps_per_second": 0.915, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.05468963631391851, |
|
"grad_norm": 1.452701449394226, |
|
"learning_rate": 0.0002997539364173515, |
|
"loss": 0.5313, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05468963631391851, |
|
"eval_loss": 0.7013862729072571, |
|
"eval_runtime": 1609.7654, |
|
"eval_samples_per_second": 29.271, |
|
"eval_steps_per_second": 0.915, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06562756357670221, |
|
"grad_norm": 2.0603904724121094, |
|
"learning_rate": 0.00029964571107821494, |
|
"loss": 0.4879, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06562756357670221, |
|
"eval_loss": 0.7164642214775085, |
|
"eval_runtime": 1609.8109, |
|
"eval_samples_per_second": 29.271, |
|
"eval_steps_per_second": 0.915, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.07656549083948591, |
|
"grad_norm": 1.048985242843628, |
|
"learning_rate": 0.00029951784199045534, |
|
"loss": 0.4665, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07656549083948591, |
|
"eval_loss": 0.7070000171661377, |
|
"eval_runtime": 1608.8038, |
|
"eval_samples_per_second": 29.289, |
|
"eval_steps_per_second": 0.916, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.08750341810226962, |
|
"grad_norm": 0.8847436904907227, |
|
"learning_rate": 0.0002993703459392396, |
|
"loss": 0.4509, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08750341810226962, |
|
"eval_loss": 0.7005062699317932, |
|
"eval_runtime": 1608.5117, |
|
"eval_samples_per_second": 29.294, |
|
"eval_steps_per_second": 0.916, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.09844134536505332, |
|
"grad_norm": 1.09526789188385, |
|
"learning_rate": 0.00029920324228613376, |
|
"loss": 0.4542, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09844134536505332, |
|
"eval_loss": 0.6932889223098755, |
|
"eval_runtime": 1608.5476, |
|
"eval_samples_per_second": 29.294, |
|
"eval_steps_per_second": 0.916, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.10937927262783702, |
|
"grad_norm": 0.9263470768928528, |
|
"learning_rate": 0.0002990165529665622, |
|
"loss": 0.4188, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.10937927262783702, |
|
"eval_loss": 0.65990149974823, |
|
"eval_runtime": 1607.8192, |
|
"eval_samples_per_second": 29.307, |
|
"eval_steps_per_second": 0.916, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.12031719989062073, |
|
"grad_norm": 1.2452396154403687, |
|
"learning_rate": 0.0002988103024869277, |
|
"loss": 0.4131, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.12031719989062073, |
|
"eval_loss": 0.6592887043952942, |
|
"eval_runtime": 1608.173, |
|
"eval_samples_per_second": 29.3, |
|
"eval_steps_per_second": 0.916, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.13125512715340443, |
|
"grad_norm": 0.8261873722076416, |
|
"learning_rate": 0.00029858451792139453, |
|
"loss": 0.4137, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.13125512715340443, |
|
"eval_loss": 0.6895950436592102, |
|
"eval_runtime": 1609.8507, |
|
"eval_samples_per_second": 29.27, |
|
"eval_steps_per_second": 0.915, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.14219305441618812, |
|
"grad_norm": 0.6665583848953247, |
|
"learning_rate": 0.0002983392289083346, |
|
"loss": 0.4087, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.14219305441618812, |
|
"eval_loss": 0.6930198073387146, |
|
"eval_runtime": 1610.0518, |
|
"eval_samples_per_second": 29.266, |
|
"eval_steps_per_second": 0.915, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.15313098167897182, |
|
"grad_norm": 0.841665506362915, |
|
"learning_rate": 0.0002980744676464371, |
|
"loss": 0.3968, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.15313098167897182, |
|
"eval_loss": 0.6691634058952332, |
|
"eval_runtime": 1610.3191, |
|
"eval_samples_per_second": 29.261, |
|
"eval_steps_per_second": 0.915, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.16406890894175555, |
|
"grad_norm": 1.1116673946380615, |
|
"learning_rate": 0.0002977902688904813, |
|
"loss": 0.4048, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.16406890894175555, |
|
"eval_loss": 0.6872161030769348, |
|
"eval_runtime": 1608.6952, |
|
"eval_samples_per_second": 29.291, |
|
"eval_steps_per_second": 0.916, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.17500683620453925, |
|
"grad_norm": 0.7878606915473938, |
|
"learning_rate": 0.00029748666994677467, |
|
"loss": 0.3953, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.17500683620453925, |
|
"eval_loss": 0.6938689351081848, |
|
"eval_runtime": 1607.7411, |
|
"eval_samples_per_second": 29.308, |
|
"eval_steps_per_second": 0.916, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.17500683620453925, |
|
"step": 160, |
|
"total_flos": 2.934122540040192e+17, |
|
"train_loss": 0.639202207326889, |
|
"train_runtime": 31420.6578, |
|
"train_samples_per_second": 22.346, |
|
"train_steps_per_second": 0.087 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2742, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 10, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.934122540040192e+17, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|