{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.412142038345337, "min": 1.412142038345337, "max": 1.4317421913146973, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 71585.71875, "min": 68603.2890625, "max": 76675.359375, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 95.33528265107212, "min": 92.94866920152091, "max": 380.2706766917293, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 48907.0, "min": 48891.0, "max": 50576.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999948.0, "min": 49959.0, "max": 1999948.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999948.0, "min": 49959.0, "max": 1999948.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.3194429874420166, "min": 0.11772950738668442, "max": 2.370976448059082, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1189.874267578125, "min": 15.540294647216797, "max": 1224.33447265625, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.6215291307915956, "min": 1.8079163441152284, "max": 3.7825915438157542, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 1857.8444440960884, "min": 238.64495742321014, "max": 1901.0664147138596, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.6215291307915956, "min": 1.8079163441152284, "max": 3.7825915438157542, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 1857.8444440960884, "min": 238.64495742321014, "max": 1901.0664147138596, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.015424298531434033, "min": 0.013561218941483983, "max": 0.01966511723100363, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.030848597062868067, "min": 0.028261511194674918, "max": 0.057302239333512264, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.04858299531042576, "min": 0.022533917302886647, "max": 0.06543632153835562, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.09716599062085152, "min": 0.045067834605773294, "max": 0.19630896461506686, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 4.440398519900002e-06, "min": 4.440398519900002e-06, "max": 0.00029529165156944997, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 8.880797039800005e-06, "min": 8.880797039800005e-06, "max": 0.0008440779186406998, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10148010000000002, "min": 0.10148010000000002, "max": 0.19843054999999998, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.20296020000000004, "min": 0.20296020000000004, "max": 0.5813593, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 8.385699000000007e-05, "min": 8.385699000000007e-05, "max": 0.004921684445000001, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00016771398000000015, "min": 0.00016771398000000015, "max": 0.01406982907, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1679263851", "python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1679266238" }, "total": 2387.442661235, "count": 1, "self": 0.49167718199987576, "children": { "run_training.setup": { "total": 0.10378191799998149, "count": 1, "self": 0.10378191799998149 }, "TrainerController.start_learning": { "total": 2386.847202135, "count": 1, "self": 4.107565909950154, "children": { "TrainerController._reset_env": { "total": 9.299930148999977, "count": 1, "self": 9.299930148999977 }, "TrainerController.advance": { "total": 2373.322413891049, "count": 231059, "self": 4.596470673002841, "children": { "env_step": { "total": 1858.1653720520003, "count": 231059, "self": 1562.115179380109, "children": { "SubprocessEnvManager._take_step": { "total": 293.17368769594856, "count": 231059, "self": 18.044804234921685, "children": { "TorchPolicy.evaluate": { "total": 275.1288834610269, "count": 222988, "self": 275.1288834610269 } } }, "workers": { "total": 2.8765049759426233, "count": 231059, "self": 0.0, "children": { "worker_root": { "total": 2378.6356424740275, "count": 231059, "is_parallel": true, "self": 1109.043328631005, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.000988613999993504, "count": 1, "is_parallel": true, "self": 0.0003060409999307012, "children": { "_process_rank_one_or_two_observation": { "total": 0.0006825730000628027, "count": 2, "is_parallel": true, "self": 0.0006825730000628027 } } }, "UnityEnvironment.step": { "total": 0.030792609000002358, "count": 1, "is_parallel": true, "self": 0.0003596500000071501, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0002318260000038208, "count": 1, "is_parallel": true, "self": 0.0002318260000038208 }, "communicator.exchange": { "total": 0.029432615000018814, "count": 1, "is_parallel": true, "self": 0.029432615000018814 }, "steps_from_proto": { "total": 0.0007685179999725733, "count": 1, "is_parallel": true, "self": 0.00026200900003914285, "children": { "_process_rank_one_or_two_observation": { "total": 0.0005065089999334305, "count": 2, "is_parallel": true, "self": 0.0005065089999334305 } } } } } } }, "UnityEnvironment.step": { "total": 1269.5923138430226, "count": 231058, "is_parallel": true, "self": 38.118532497952856, "children": { "UnityEnvironment._generate_step_input": { "total": 80.02550050400464, "count": 231058, "is_parallel": true, "self": 80.02550050400464 }, "communicator.exchange": { "total": 1061.2737338630347, "count": 231058, "is_parallel": true, "self": 1061.2737338630347 }, "steps_from_proto": { "total": 90.17454697803021, "count": 231058, "is_parallel": true, "self": 36.03128876496993, "children": { "_process_rank_one_or_two_observation": { "total": 54.14325821306028, "count": 462116, "is_parallel": true, "self": 54.14325821306028 } } } } } } } } } } }, "trainer_advance": { "total": 510.56057116604575, "count": 231059, "self": 6.5054286510249995, "children": { "process_trajectory": { "total": 135.9607855190198, "count": 231059, "self": 134.60456597201943, "children": { "RLTrainer._checkpoint": { "total": 1.3562195470003644, "count": 10, "self": 1.3562195470003644 } } }, "_update_policy": { "total": 368.09435699600095, "count": 96, "self": 310.841100506, "children": { "TorchPPOOptimizer.update": { "total": 57.25325649000098, "count": 2880, "self": 57.25325649000098 } } } } } } }, "trainer_threads": { "total": 1.2990003597224131e-06, "count": 1, "self": 1.2990003597224131e-06 }, "TrainerController._save_models": { "total": 0.117290886000319, "count": 1, "self": 0.0029260330002216506, "children": { "RLTrainer._checkpoint": { "total": 0.11436485300009736, "count": 1, "self": 0.11436485300009736 } } } } } } }