ppo-Huggy / run_logs /timers.json
Andyrasika's picture
Huggy
01f4403
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4049053192138672,
"min": 1.4049053192138672,
"max": 1.4279431104660034,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 68261.5390625,
"min": 68261.5390625,
"max": 77998.875,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 92.61235955056179,
"min": 80.80360065466448,
"max": 405.53225806451616,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49455.0,
"min": 48772.0,
"max": 50286.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999947.0,
"min": 49676.0,
"max": 1999947.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999947.0,
"min": 49676.0,
"max": 1999947.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.445343494415283,
"min": 0.10563024133443832,
"max": 2.491300344467163,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1305.8134765625,
"min": 12.99251937866211,
"max": 1482.931884765625,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.812229439783632,
"min": 1.8230786844482267,
"max": 4.002624439565759,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2035.7305208444595,
"min": 224.23867818713188,
"max": 2408.0704147815704,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.812229439783632,
"min": 1.8230786844482267,
"max": 4.002624439565759,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2035.7305208444595,
"min": 224.23867818713188,
"max": 2408.0704147815704,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.015010151500933313,
"min": 0.012864824100688566,
"max": 0.020264136506981837,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.04503045450279994,
"min": 0.02572964820137713,
"max": 0.060792409520945515,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.05266350222130617,
"min": 0.02205253584931294,
"max": 0.06283031271563636,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.1579905066639185,
"min": 0.04410507169862588,
"max": 0.18849093814690907,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.590598803166675e-06,
"min": 3.590598803166675e-06,
"max": 0.000295355476548175,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.0771796409500025e-05,
"min": 1.0771796409500025e-05,
"max": 0.0008442535685821498,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10119683333333336,
"min": 0.10119683333333336,
"max": 0.19845182499999997,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.3035905000000001,
"min": 0.20760785000000004,
"max": 0.5814178499999999,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.972198333333347e-05,
"min": 6.972198333333347e-05,
"max": 0.004922746067500002,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0002091659500000004,
"min": 0.0002091659500000004,
"max": 0.014072750714999999,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1692535484",
"python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1692537355"
},
"total": 1870.30333868,
"count": 1,
"self": 0.27480599800014716,
"children": {
"run_training.setup": {
"total": 0.038165552000009484,
"count": 1,
"self": 0.038165552000009484
},
"TrainerController.start_learning": {
"total": 1869.9903671299999,
"count": 1,
"self": 3.6400500439331154,
"children": {
"TrainerController._reset_env": {
"total": 3.8407117499999686,
"count": 1,
"self": 3.8407117499999686
},
"TrainerController.advance": {
"total": 1862.3979072970671,
"count": 232096,
"self": 3.793431612032464,
"children": {
"env_step": {
"total": 1416.1478082070776,
"count": 232096,
"self": 1155.0616105500753,
"children": {
"SubprocessEnvManager._take_step": {
"total": 258.6500575229903,
"count": 232096,
"self": 13.978300056017929,
"children": {
"TorchPolicy.evaluate": {
"total": 244.6717574669724,
"count": 222851,
"self": 244.6717574669724
}
}
},
"workers": {
"total": 2.436140134012021,
"count": 232096,
"self": 0.0,
"children": {
"worker_root": {
"total": 1863.2781108430574,
"count": 232096,
"is_parallel": true,
"self": 929.4247217811152,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0007167669999716964,
"count": 1,
"is_parallel": true,
"self": 0.000237169999991238,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00047959699998045835,
"count": 2,
"is_parallel": true,
"self": 0.00047959699998045835
}
}
},
"UnityEnvironment.step": {
"total": 0.02023059199996169,
"count": 1,
"is_parallel": true,
"self": 0.0002246870000135459,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00016283899992686202,
"count": 1,
"is_parallel": true,
"self": 0.00016283899992686202
},
"communicator.exchange": {
"total": 0.019409296000048926,
"count": 1,
"is_parallel": true,
"self": 0.019409296000048926
},
"steps_from_proto": {
"total": 0.00043376999997235544,
"count": 1,
"is_parallel": true,
"self": 0.00012681899988820078,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00030695100008415466,
"count": 2,
"is_parallel": true,
"self": 0.00030695100008415466
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 933.8533890619423,
"count": 232095,
"is_parallel": true,
"self": 26.4780176498532,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 47.71850990507221,
"count": 232095,
"is_parallel": true,
"self": 47.71850990507221
},
"communicator.exchange": {
"total": 799.0823164670242,
"count": 232095,
"is_parallel": true,
"self": 799.0823164670242
},
"steps_from_proto": {
"total": 60.5745450399927,
"count": 232095,
"is_parallel": true,
"self": 21.354120618987395,
"children": {
"_process_rank_one_or_two_observation": {
"total": 39.220424421005305,
"count": 464190,
"is_parallel": true,
"self": 39.220424421005305
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 442.456667477957,
"count": 232096,
"self": 5.461644060037884,
"children": {
"process_trajectory": {
"total": 113.83662973991954,
"count": 232096,
"self": 112.67993396691963,
"children": {
"RLTrainer._checkpoint": {
"total": 1.1566957729999103,
"count": 10,
"self": 1.1566957729999103
}
}
},
"_update_policy": {
"total": 323.1583936779996,
"count": 97,
"self": 279.33159016500065,
"children": {
"TorchPPOOptimizer.update": {
"total": 43.826803512998936,
"count": 2910,
"self": 43.826803512998936
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.0069998097606003e-06,
"count": 1,
"self": 1.0069998097606003e-06
},
"TrainerController._save_models": {
"total": 0.11169703199993819,
"count": 1,
"self": 0.0017893429999276123,
"children": {
"RLTrainer._checkpoint": {
"total": 0.10990768900001058,
"count": 1,
"self": 0.10990768900001058
}
}
}
}
}
}
}