{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4048871994018555, "min": 1.4048871994018555, "max": 1.4278028011322021, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 70147.421875, "min": 68635.7265625, "max": 78814.484375, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 86.07731958762886, "min": 85.68804159445408, "max": 410.0650406504065, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 50097.0, "min": 48720.0, "max": 50438.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999992.0, "min": 49812.0, "max": 1999992.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999992.0, "min": 49812.0, "max": 1999992.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4547152519226074, "min": 0.036121074110269547, "max": 2.508781671524048, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1428.644287109375, "min": 4.406771183013916, "max": 1428.644287109375, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.7428781648272094, "min": 1.6429010960655135, "max": 3.927283979234872, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2178.3550919294357, "min": 200.43393371999264, "max": 2178.3550919294357, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.7428781648272094, "min": 1.6429010960655135, "max": 3.927283979234872, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2178.3550919294357, "min": 200.43393371999264, "max": 2178.3550919294357, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.016107808962018074, "min": 0.01337862528711816, "max": 0.020056149534260236, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.048323426886054226, "min": 0.027667107045514665, "max": 0.05602611232801186, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.05747392972310384, "min": 0.023587215691804886, "max": 0.059679197085400426, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.17242178916931153, "min": 0.04717443138360977, "max": 0.1759285061309735, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.4479488507166627e-06, "min": 3.4479488507166627e-06, "max": 0.00029536912654362494, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.0343846552149989e-05, "min": 1.0343846552149989e-05, "max": 0.0008444085185305001, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10114928333333334, "min": 0.10114928333333334, "max": 0.19845637500000002, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30344785, "min": 0.20742345000000006, "max": 0.5814695000000001, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 6.734923833333328e-05, "min": 6.734923833333328e-05, "max": 0.0049229731125, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00020204771499999984, "min": 0.00020204771499999984, "max": 0.01407532805, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1694725839", "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1694728349" }, "total": 2510.180224487, "count": 1, "self": 0.4384881669998322, "children": { "run_training.setup": { "total": 0.0688454380000394, "count": 1, "self": 0.0688454380000394 }, "TrainerController.start_learning": { "total": 2509.672890882, "count": 1, "self": 4.772162512032537, "children": { "TrainerController._reset_env": { "total": 4.086899592000009, "count": 1, "self": 4.086899592000009 }, "TrainerController.advance": { "total": 2500.692519578967, "count": 232079, "self": 4.944938914058184, "children": { "env_step": { "total": 1944.931057575993, "count": 232079, "self": 1638.395153172137, "children": { "SubprocessEnvManager._take_step": { "total": 303.51147907987547, "count": 232079, "self": 16.96425078986408, "children": { "TorchPolicy.evaluate": { "total": 286.5472282900114, "count": 223013, "self": 286.5472282900114 } } }, "workers": { "total": 3.024425323980495, "count": 232079, "self": 0.0, "children": { "worker_root": { "total": 2501.983143351965, "count": 232079, "is_parallel": true, "self": 1165.4596545018157, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0009299309999732941, "count": 1, "is_parallel": true, "self": 0.0002486360000375498, "children": { "_process_rank_one_or_two_observation": { "total": 0.0006812949999357443, "count": 2, "is_parallel": true, "self": 0.0006812949999357443 } } }, "UnityEnvironment.step": { "total": 0.03039059599996108, "count": 1, "is_parallel": true, "self": 0.00032073599982140877, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00024305800002366595, "count": 1, "is_parallel": true, "self": 0.00024305800002366595 }, "communicator.exchange": { "total": 0.029104368000048453, "count": 1, "is_parallel": true, "self": 0.029104368000048453 }, "steps_from_proto": { "total": 0.0007224340000675511, "count": 1, "is_parallel": true, "self": 0.00021517700020012853, "children": { "_process_rank_one_or_two_observation": { "total": 0.0005072569998674226, "count": 2, "is_parallel": true, "self": 0.0005072569998674226 } } } } } } }, "UnityEnvironment.step": { "total": 1336.523488850149, "count": 232078, "is_parallel": true, "self": 40.630606590129446, "children": { "UnityEnvironment._generate_step_input": { "total": 86.32252719198004, "count": 232078, "is_parallel": true, "self": 86.32252719198004 }, "communicator.exchange": { "total": 1108.9368541230401, "count": 232078, "is_parallel": true, "self": 1108.9368541230401 }, "steps_from_proto": { "total": 100.63350094499947, "count": 232078, "is_parallel": true, "self": 38.06560514001069, "children": { "_process_rank_one_or_two_observation": { "total": 62.56789580498878, "count": 464156, "is_parallel": true, "self": 62.56789580498878 } } } } } } } } } } }, "trainer_advance": { "total": 550.8165230889163, "count": 232079, "self": 6.649744141836891, "children": { "process_trajectory": { "total": 143.8730274590805, "count": 232079, "self": 142.3843067040807, "children": { "RLTrainer._checkpoint": { "total": 1.4887207549998038, "count": 10, "self": 1.4887207549998038 } } }, "_update_policy": { "total": 400.29375148799886, "count": 97, "self": 340.45591408399423, "children": { "TorchPPOOptimizer.update": { "total": 59.83783740400463, "count": 2910, "self": 59.83783740400463 } } } } } } }, "trainer_threads": { "total": 8.570000318286475e-07, "count": 1, "self": 8.570000318286475e-07 }, "TrainerController._save_models": { "total": 0.12130834200024765, "count": 1, "self": 0.0020043520003127924, "children": { "RLTrainer._checkpoint": { "total": 0.11930398999993486, "count": 1, "self": 0.11930398999993486 } } } } } } }