{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4186748266220093, "min": 1.4186748266220093, "max": 1.4265297651290894, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 70864.2265625, "min": 68991.1171875, "max": 76873.75, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 83.95238095238095, "min": 73.75748502994011, "max": 410.0409836065574, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49364.0, "min": 49073.0, "max": 50357.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999902.0, "min": 49806.0, "max": 1999902.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999902.0, "min": 49806.0, "max": 1999902.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.451608180999756, "min": 0.05042506009340286, "max": 2.509305953979492, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1441.545654296875, "min": 6.1014323234558105, "max": 1636.81005859375, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.7875565440476344, "min": 1.9085291205358899, "max": 4.01449012031922, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2227.083247900009, "min": 230.93202358484268, "max": 2642.61474275589, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.7875565440476344, "min": 1.9085291205358899, "max": 4.01449012031922, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2227.083247900009, "min": 230.93202358484268, "max": 2642.61474275589, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.01629057851532707, "min": 0.013970197132221075, "max": 0.019364503834367497, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.04887173554598121, "min": 0.02794039426444215, "max": 0.05569177705074253, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.0660530130378902, "min": 0.02126063934992999, "max": 0.0660530130378902, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.1981590391136706, "min": 0.04252127869985998, "max": 0.1981590391136706, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 1.2369654297999991e-06, "min": 1.2369654297999991e-06, "max": 9.845272654727499e-05, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 3.7108962893999977e-06, "min": 3.7108962893999977e-06, "max": 0.0002813238186762, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10123686666666669, "min": 0.10123686666666669, "max": 0.198452725, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30371060000000005, "min": 0.20761845, "max": 0.5813238000000001, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 7.048277999999995e-05, "min": 7.048277999999995e-05, "max": 0.0048243382525, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00021144833999999987, "min": 0.00021144833999999987, "max": 0.01378673382, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1691338053", "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1691343873" }, "total": 5819.913077859001, "count": 1, "self": 0.6019856820003042, "children": { "run_training.setup": { "total": 0.05611564700006966, "count": 1, "self": 0.05611564700006966 }, "TrainerController.start_learning": { "total": 5819.25497653, "count": 1, "self": 7.979269543033297, "children": { "TrainerController._reset_env": { "total": 3.256375319999961, "count": 1, "self": 3.256375319999961 }, "TrainerController.advance": { "total": 5807.850546047966, "count": 232964, "self": 8.767589436137314, "children": { "env_step": { "total": 2974.428881672819, "count": 232964, "self": 2458.3139398113854, "children": { "SubprocessEnvManager._take_step": { "total": 510.70896391718463, "count": 232964, "self": 30.011927683183103, "children": { "TorchPolicy.evaluate": { "total": 480.69703623400153, "count": 223010, "self": 480.69703623400153 } } }, "workers": { "total": 5.405977944249003, "count": 232964, "self": 0.0, "children": { "worker_root": { "total": 5805.05493377517, "count": 232964, "is_parallel": true, "self": 3835.0637592823227, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0010457320000796244, "count": 1, "is_parallel": true, "self": 0.00029718900020725414, "children": { "_process_rank_one_or_two_observation": { "total": 0.0007485429998723703, "count": 2, "is_parallel": true, "self": 0.0007485429998723703 } } }, "UnityEnvironment.step": { "total": 0.03758028499999, "count": 1, "is_parallel": true, "self": 0.00042384800008221646, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00028940599997895333, "count": 1, "is_parallel": true, "self": 0.00028940599997895333 }, "communicator.exchange": { "total": 0.0359255219999568, "count": 1, "is_parallel": true, "self": 0.0359255219999568 }, "steps_from_proto": { "total": 0.0009415089999720294, "count": 1, "is_parallel": true, "self": 0.00024081599997316516, "children": { "_process_rank_one_or_two_observation": { "total": 0.0007006929999988643, "count": 2, "is_parallel": true, "self": 0.0007006929999988643 } } } } } } }, "UnityEnvironment.step": { "total": 1969.9911744928472, "count": 232963, "is_parallel": true, "self": 60.57097024876521, "children": { "UnityEnvironment._generate_step_input": { "total": 88.73637465901652, "count": 232963, "is_parallel": true, "self": 88.73637465901652 }, "communicator.exchange": { "total": 1676.6948118630662, "count": 232963, "is_parallel": true, "self": 1676.6948118630662 }, "steps_from_proto": { "total": 143.9890177219993, "count": 232963, "is_parallel": true, "self": 47.942985211207315, "children": { "_process_rank_one_or_two_observation": { "total": 96.04603251079197, "count": 465926, "is_parallel": true, "self": 96.04603251079197 } } } } } } } } } } }, "trainer_advance": { "total": 2824.65407493901, "count": 232964, "self": 12.819008111138373, "children": { "process_trajectory": { "total": 272.26632644786946, "count": 232964, "self": 270.15657773587077, "children": { "RLTrainer._checkpoint": { "total": 2.109748711998691, "count": 10, "self": 2.109748711998691 } } }, "_update_policy": { "total": 2539.5687403800025, "count": 97, "self": 588.3170563150077, "children": { "TorchPPOOptimizer.update": { "total": 1951.2516840649948, "count": 3880, "self": 1951.2516840649948 } } } } } } }, "trainer_threads": { "total": 1.1290003385511227e-06, "count": 1, "self": 1.1290003385511227e-06 }, "TrainerController._save_models": { "total": 0.1687844899997799, "count": 1, "self": 0.004356408999228734, "children": { "RLTrainer._checkpoint": { "total": 0.16442808100055117, "count": 1, "self": 0.16442808100055117 } } } } } } }