ppo-Huggy / run_logs /timers.json
maraoz's picture
Huggy
47e9989
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4027098417282104,
"min": 1.4027098417282104,
"max": 1.4272503852844238,
"count": 38
},
"Huggy.Policy.Entropy.sum": {
"value": 69933.5,
"min": 56408.9453125,
"max": 72006.1171875,
"count": 38
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 76.77449455676516,
"min": 75.50995405819296,
"max": 284.2480620155039,
"count": 38
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49366.0,
"min": 36668.0,
"max": 49668.0,
"count": 38
},
"Huggy.Step.mean": {
"value": 1999915.0,
"min": 149837.0,
"max": 1999915.0,
"count": 38
},
"Huggy.Step.sum": {
"value": 1999915.0,
"min": 149837.0,
"max": 1999915.0,
"count": 38
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4969820976257324,
"min": 0.6142274141311646,
"max": 2.5275051593780518,
"count": 38
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1605.5594482421875,
"min": 78.62110900878906,
"max": 1605.5594482421875,
"count": 38
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.798717265551791,
"min": 2.9495590571314096,
"max": 4.001001757807717,
"count": 38
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2442.5752017498016,
"min": 377.54355931282043,
"max": 2499.637472629547,
"count": 38
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.798717265551791,
"min": 2.9495590571314096,
"max": 4.001001757807717,
"count": 38
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2442.5752017498016,
"min": 377.54355931282043,
"max": 2499.637472629547,
"count": 38
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.01808087213624579,
"min": 0.012528159525148415,
"max": 0.020715488920298716,
"count": 38
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.03616174427249158,
"min": 0.018980470420016597,
"max": 0.06098752787026267,
"count": 38
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.06257759500294924,
"min": 0.021718453615903854,
"max": 0.06257759500294924,
"count": 38
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.12515519000589848,
"min": 0.021718453615903854,
"max": 0.1841054182499647,
"count": 38
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.850073716674988e-06,
"min": 3.850073716674988e-06,
"max": 0.00027985230671589994,
"count": 38
},
"Huggy.Policy.LearningRate.sum": {
"value": 7.700147433349975e-06,
"min": 7.700147433349975e-06,
"max": 0.0008210698763100501,
"count": 38
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10128332500000001,
"min": 0.10128332500000001,
"max": 0.19328409999999993,
"count": 38
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20256665000000001,
"min": 0.19328409999999993,
"max": 0.5736899499999999,
"count": 38
},
"Huggy.Policy.Beta.mean": {
"value": 7.403791749999983e-05,
"min": 7.403791749999983e-05,
"max": 0.004664876590000001,
"count": 38
},
"Huggy.Policy.Beta.sum": {
"value": 0.00014807583499999965,
"min": 0.00014807583499999965,
"max": 0.013687128505,
"count": 38
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 38
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 38
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1675384406",
"python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn --resume ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1675386548"
},
"total": 2141.902958832,
"count": 1,
"self": 0.43451878399991983,
"children": {
"run_training.setup": {
"total": 0.10099393900003406,
"count": 1,
"self": 0.10099393900003406
},
"TrainerController.start_learning": {
"total": 2141.367446109,
"count": 1,
"self": 3.931397098968773,
"children": {
"TrainerController._reset_env": {
"total": 5.8351429239999675,
"count": 1,
"self": 5.8351429239999675
},
"TrainerController.advance": {
"total": 2131.490610933031,
"count": 220370,
"self": 4.170159091981077,
"children": {
"env_step": {
"total": 1690.9400327430474,
"count": 220370,
"self": 1421.8756554991332,
"children": {
"SubprocessEnvManager._take_step": {
"total": 266.53078453299133,
"count": 220370,
"self": 13.89158923398952,
"children": {
"TorchPolicy.evaluate": {
"total": 252.6391952990018,
"count": 210309,
"self": 62.26420008105629,
"children": {
"TorchPolicy.sample_actions": {
"total": 190.37499521794552,
"count": 210309,
"self": 190.37499521794552
}
}
}
}
},
"workers": {
"total": 2.5335927109229033,
"count": 220370,
"self": 0.0,
"children": {
"worker_root": {
"total": 2133.4522089779316,
"count": 220370,
"is_parallel": true,
"self": 960.2063948889408,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0008028959999819563,
"count": 1,
"is_parallel": true,
"self": 0.0003002929999524895,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005026030000294668,
"count": 2,
"is_parallel": true,
"self": 0.0005026030000294668
}
}
},
"UnityEnvironment.step": {
"total": 0.027544895000005454,
"count": 1,
"is_parallel": true,
"self": 0.00029913000014403224,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00018594699997720454,
"count": 1,
"is_parallel": true,
"self": 0.00018594699997720454
},
"communicator.exchange": {
"total": 0.026118452999980946,
"count": 1,
"is_parallel": true,
"self": 0.026118452999980946
},
"steps_from_proto": {
"total": 0.0009413649999032714,
"count": 1,
"is_parallel": true,
"self": 0.0002550019999034703,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006863629999998011,
"count": 2,
"is_parallel": true,
"self": 0.0006863629999998011
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1173.2458140889908,
"count": 220369,
"is_parallel": true,
"self": 33.08206983395803,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 74.16290157603316,
"count": 220369,
"is_parallel": true,
"self": 74.16290157603316
},
"communicator.exchange": {
"total": 973.6695953199352,
"count": 220369,
"is_parallel": true,
"self": 973.6695953199352
},
"steps_from_proto": {
"total": 92.33124735906438,
"count": 220369,
"is_parallel": true,
"self": 38.697783664988265,
"children": {
"_process_rank_one_or_two_observation": {
"total": 53.63346369407611,
"count": 440738,
"is_parallel": true,
"self": 53.63346369407611
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 436.38041909800256,
"count": 220370,
"self": 5.844339883985185,
"children": {
"process_trajectory": {
"total": 147.5605605180166,
"count": 220370,
"self": 146.47950046801577,
"children": {
"RLTrainer._checkpoint": {
"total": 1.081060050000815,
"count": 10,
"self": 1.081060050000815
}
}
},
"_update_policy": {
"total": 282.9755186960008,
"count": 91,
"self": 233.61468696299585,
"children": {
"TorchPPOOptimizer.update": {
"total": 49.36083173300494,
"count": 2730,
"self": 49.36083173300494
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.330001375928987e-07,
"count": 1,
"self": 9.330001375928987e-07
},
"TrainerController._save_models": {
"total": 0.11029422000001432,
"count": 1,
"self": 0.002647132999754831,
"children": {
"RLTrainer._checkpoint": {
"total": 0.10764708700025949,
"count": 1,
"self": 0.10764708700025949
}
}
}
}
}
}
}