{ "name": "root", "gauges": { "SoccerTwos.Policy.Entropy.mean": { "value": 1.8188538551330566, "min": 1.781403660774231, "max": 3.2958128452301025, "count": 1000 }, "SoccerTwos.Policy.Entropy.sum": { "value": 36551.6875, "min": 24124.41015625, "max": 129180.15625, "count": 1000 }, "SoccerTwos.Environment.EpisodeLength.mean": { "value": 68.12328767123287, "min": 44.27272727272727, "max": 999.0, "count": 1000 }, "SoccerTwos.Environment.EpisodeLength.sum": { "value": 19892.0, "min": 12748.0, "max": 27952.0, "count": 1000 }, "SoccerTwos.Self-play.ELO.mean": { "value": 1482.6114705511727, "min": 1195.030121187817, "max": 1551.3509553660122, "count": 915 }, "SoccerTwos.Self-play.ELO.sum": { "value": 216461.2747004712, "min": 2390.060242375634, "max": 327571.1699347649, "count": 915 }, "SoccerTwos.Step.mean": { "value": 9999982.0, "min": 9992.0, "max": 9999982.0, "count": 1000 }, "SoccerTwos.Step.sum": { "value": 9999982.0, "min": 9992.0, "max": 9999982.0, "count": 1000 }, "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": { "value": -0.03705725073814392, "min": -0.13788607716560364, "max": 0.11721375584602356, "count": 1000 }, "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": { "value": -5.410358428955078, "min": -18.06307601928711, "max": 17.642162322998047, "count": 1000 }, "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": { "value": -0.03660492226481438, "min": -0.14178521931171417, "max": 0.11586101353168488, "count": 1000 }, "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": { "value": -5.344318389892578, "min": -18.573863983154297, "max": 17.03980255126953, "count": 1000 }, "SoccerTwos.Environment.CumulativeReward.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 1000 }, "SoccerTwos.Environment.CumulativeReward.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 1000 }, "SoccerTwos.Policy.ExtrinsicReward.mean": { "value": -0.05301643969261483, "min": -0.6666666666666666, "max": 0.36468837150307587, "count": 1000 }, "SoccerTwos.Policy.ExtrinsicReward.sum": { "value": -7.740400195121765, "min": -58.76360011100769, "max": 56.248599886894226, "count": 1000 }, "SoccerTwos.Environment.GroupCumulativeReward.mean": { "value": -0.05301643969261483, "min": -0.6666666666666666, "max": 0.36468837150307587, "count": 1000 }, "SoccerTwos.Environment.GroupCumulativeReward.sum": { "value": -7.740400195121765, "min": -58.76360011100769, "max": 56.248599886894226, "count": 1000 }, "SoccerTwos.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 1000 }, "SoccerTwos.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 1000 }, "SoccerTwos.Losses.PolicyLoss.mean": { "value": 0.015276352792473821, "min": 0.009830443564017818, "max": 0.025515147866826737, "count": 593 }, "SoccerTwos.Losses.PolicyLoss.sum": { "value": 0.015276352792473821, "min": 0.009830443564017818, "max": 0.025515147866826737, "count": 593 }, "SoccerTwos.Losses.ValueLoss.mean": { "value": 0.10177272527168195, "min": 2.6980898759632055e-07, "max": 0.1258347670858105, "count": 593 }, "SoccerTwos.Losses.ValueLoss.sum": { "value": 0.10177272527168195, "min": 2.6980898759632055e-07, "max": 0.1258347670858105, "count": 593 }, "SoccerTwos.Losses.BaselineLoss.mean": { "value": 0.10286407048503558, "min": 2.608114281580735e-07, "max": 0.12716998159885406, "count": 593 }, "SoccerTwos.Losses.BaselineLoss.sum": { "value": 0.10286407048503558, "min": 2.608114281580735e-07, "max": 0.12716998159885406, "count": 593 }, "SoccerTwos.Policy.LearningRate.mean": { "value": 0.0003, "min": 0.0003, "max": 0.0003, "count": 593 }, "SoccerTwos.Policy.LearningRate.sum": { "value": 0.0003, "min": 0.0003, "max": 0.0003, "count": 593 }, "SoccerTwos.Policy.Epsilon.mean": { "value": 0.20000000000000004, "min": 0.20000000000000004, "max": 0.20000000000000004, "count": 593 }, "SoccerTwos.Policy.Epsilon.sum": { "value": 0.20000000000000004, "min": 0.20000000000000004, "max": 0.20000000000000004, "count": 593 }, "SoccerTwos.Policy.Beta.mean": { "value": 0.005, "min": 0.005, "max": 0.005, "count": 593 }, "SoccerTwos.Policy.Beta.sum": { "value": 0.005, "min": 0.005, "max": 0.005, "count": 593 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1676196126", "python_version": "3.9.5 (default, Nov 23 2021, 15:27:38) \n[GCC 9.3.0]", "command_line_arguments": "/home/akghxhs55/.local/share/virtualenvs/unit7-nHLSdMlF/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos-2 --no-graphics --force", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu115", "numpy_version": "1.21.2", "end_time_seconds": "1676207973" }, "total": 11847.087752068, "count": 1, "self": 0.21812908399806474, "children": { "run_training.setup": { "total": 0.006077324000216322, "count": 1, "self": 0.006077324000216322 }, "TrainerController.start_learning": { "total": 11846.863545660002, "count": 1, "self": 9.1213390672674, "children": { "TrainerController._reset_env": { "total": 3.7617044860235183, "count": 50, "self": 3.7617044860235183 }, "TrainerController.advance": { "total": 11833.799101046712, "count": 672042, "self": 9.104758377132384, "children": { "env_step": { "total": 9662.090405053012, "count": 672042, "self": 6407.503559412209, "children": { "SubprocessEnvManager._take_step": { "total": 3248.586183221316, "count": 672042, "self": 56.308803162468394, "children": { "TorchPolicy.evaluate": { "total": 3192.2773800588475, "count": 1270172, "self": 3192.2773800588475 } } }, "workers": { "total": 6.000662419486616, "count": 672042, "self": 0.0, "children": { "worker_root": { "total": 11826.591512960797, "count": 672042, "is_parallel": true, "self": 6303.622571555945, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0027747630010708235, "count": 2, "is_parallel": true, "self": 0.0008088170034170616, "children": { "_process_rank_one_or_two_observation": { "total": 0.001965945997653762, "count": 8, "is_parallel": true, "self": 0.001965945997653762 } } }, "UnityEnvironment.step": { "total": 0.02019324800130562, "count": 1, "is_parallel": true, "self": 0.0005006530009268317, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0003593780020310078, "count": 1, "is_parallel": true, "self": 0.0003593780020310078 }, "communicator.exchange": { "total": 0.018213053997897077, "count": 1, "is_parallel": true, "self": 0.018213053997897077 }, "steps_from_proto": { "total": 0.0011201630004507024, "count": 2, "is_parallel": true, "self": 0.0002734020017669536, "children": { "_process_rank_one_or_two_observation": { "total": 0.0008467609986837488, "count": 8, "is_parallel": true, "self": 0.0008467609986837488 } } } } } } }, "UnityEnvironment.step": { "total": 5522.931100370846, "count": 672041, "is_parallel": true, "self": 185.67616058178828, "children": { "UnityEnvironment._generate_step_input": { "total": 108.41729930735892, "count": 672041, "is_parallel": true, "self": 108.41729930735892 }, "communicator.exchange": { "total": 4745.452247978563, "count": 672041, "is_parallel": true, "self": 4745.452247978563 }, "steps_from_proto": { "total": 483.38539250313625, "count": 1344082, "is_parallel": true, "self": 94.26806695579944, "children": { "_process_rank_one_or_two_observation": { "total": 389.1173255473368, "count": 5376328, "is_parallel": true, "self": 389.1173255473368 } } } } }, "steps_from_proto": { "total": 0.03784103400539607, "count": 98, "is_parallel": true, "self": 0.007442553007422248, "children": { "_process_rank_one_or_two_observation": { "total": 0.03039848099797382, "count": 392, "is_parallel": true, "self": 0.03039848099797382 } } } } } } } } }, "trainer_advance": { "total": 2162.603937616568, "count": 672042, "self": 69.20696104407762, "children": { "process_trajectory": { "total": 943.1222971013922, "count": 672042, "self": 939.9456614824012, "children": { "RLTrainer._checkpoint": { "total": 3.1766356189909857, "count": 20, "self": 3.1766356189909857 } } }, "_update_policy": { "total": 1150.2746794710984, "count": 593, "self": 628.9028347421445, "children": { "TorchPOCAOptimizer.update": { "total": 521.3718447289539, "count": 14232, "self": 521.3718447289539 } } } } } } }, "trainer_threads": { "total": 5.320034688338637e-07, "count": 1, "self": 5.320034688338637e-07 }, "TrainerController._save_models": { "total": 0.18140052799572004, "count": 1, "self": 0.005465680995257571, "children": { "RLTrainer._checkpoint": { "total": 0.17593484700046247, "count": 1, "self": 0.17593484700046247 } } } } } } }