poca-SoccerTwos / run_logs /timers.json
pietroorlandi's picture
First push sjfdj
913faae verified
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.8330649137496948,
"min": 1.8330649137496948,
"max": 3.2957115173339844,
"count": 500
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 33963.02734375,
"min": 15502.99609375,
"max": 129344.234375,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 48.7979797979798,
"min": 38.31782945736434,
"max": 999.0,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19324.0,
"min": 9592.0,
"max": 30748.0,
"count": 500
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1524.4819392147065,
"min": 1187.9012021155336,
"max": 1524.6000884235914,
"count": 480
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 301847.4239645119,
"min": 2375.802404231067,
"max": 365472.21019109664,
"count": 480
},
"SoccerTwos.Step.mean": {
"value": 4999982.0,
"min": 9392.0,
"max": 4999982.0,
"count": 500
},
"SoccerTwos.Step.sum": {
"value": 4999982.0,
"min": 9392.0,
"max": 4999982.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.012674360536038876,
"min": -0.1265983283519745,
"max": 0.17545248568058014,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 2.522197723388672,
"min": -26.369752883911133,
"max": 33.6868782043457,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.01138981617987156,
"min": -0.12557150423526764,
"max": 0.18099145591259003,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 2.266573429107666,
"min": -25.396196365356445,
"max": 34.19306182861328,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.02168442316390761,
"min": -0.5441052631327981,
"max": 0.69062413789075,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -4.315200209617615,
"min": -52.113200068473816,
"max": 65.66200083494186,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.02168442316390761,
"min": -0.5441052631327981,
"max": 0.69062413789075,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -4.315200209617615,
"min": -52.113200068473816,
"max": 65.66200083494186,
"count": 500
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.01662309595073263,
"min": 0.009347411760730513,
"max": 0.022080293402541428,
"count": 239
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.01662309595073263,
"min": 0.009347411760730513,
"max": 0.022080293402541428,
"count": 239
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.11728067398071289,
"min": 6.219933129614219e-05,
"max": 0.12279844060540199,
"count": 239
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.11728067398071289,
"min": 6.219933129614219e-05,
"max": 0.12279844060540199,
"count": 239
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.11979527647296588,
"min": 6.267627722991164e-05,
"max": 0.12584476098418235,
"count": 239
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.11979527647296588,
"min": 6.267627722991164e-05,
"max": 0.12584476098418235,
"count": 239
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 239
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 239
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 239
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 239
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 239
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 239
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1714732509",
"python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
"command_line_arguments": "/home/pietro/Progetti/AI/007_SelfPlay_Soccer_MLAgents/.venv/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwosID1 --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.3.0+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1714742987"
},
"total": 10477.751113665001,
"count": 1,
"self": 7.197908389001896,
"children": {
"run_training.setup": {
"total": 0.01598612799989496,
"count": 1,
"self": 0.01598612799989496
},
"TrainerController.start_learning": {
"total": 10470.537219148,
"count": 1,
"self": 6.525019467309903,
"children": {
"TrainerController._reset_env": {
"total": 4.068710587997884,
"count": 25,
"self": 4.068710587997884
},
"TrainerController.advance": {
"total": 10459.649092450692,
"count": 341422,
"self": 7.184961536784613,
"children": {
"env_step": {
"total": 8299.253499821456,
"count": 341422,
"self": 5453.531355110759,
"children": {
"SubprocessEnvManager._take_step": {
"total": 2841.4049587233494,
"count": 341422,
"self": 43.64772020249029,
"children": {
"TorchPolicy.evaluate": {
"total": 2797.757238520859,
"count": 632696,
"self": 2797.757238520859
}
}
},
"workers": {
"total": 4.3171859873473295,
"count": 341422,
"self": 0.0,
"children": {
"worker_root": {
"total": 10457.179248886992,
"count": 341422,
"is_parallel": true,
"self": 5716.032563683032,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.003567404000023089,
"count": 2,
"is_parallel": true,
"self": 0.0009717480002109369,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.002595655999812152,
"count": 8,
"is_parallel": true,
"self": 0.002595655999812152
}
}
},
"UnityEnvironment.step": {
"total": 0.02980574000002889,
"count": 1,
"is_parallel": true,
"self": 0.00047191800013024476,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0003521339999679185,
"count": 1,
"is_parallel": true,
"self": 0.0003521339999679185
},
"communicator.exchange": {
"total": 0.0269140220002555,
"count": 1,
"is_parallel": true,
"self": 0.0269140220002555
},
"steps_from_proto": {
"total": 0.002067665999675228,
"count": 2,
"is_parallel": true,
"self": 0.00032129000010172604,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0017463759995735018,
"count": 8,
"is_parallel": true,
"self": 0.0017463759995735018
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 4741.106920668955,
"count": 341421,
"is_parallel": true,
"self": 147.48811536729318,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 88.87665697753619,
"count": 341421,
"is_parallel": true,
"self": 88.87665697753619
},
"communicator.exchange": {
"total": 4063.256905320186,
"count": 341421,
"is_parallel": true,
"self": 4063.256905320186
},
"steps_from_proto": {
"total": 441.48524300394,
"count": 682842,
"is_parallel": true,
"self": 93.22880822151319,
"children": {
"_process_rank_one_or_two_observation": {
"total": 348.2564347824268,
"count": 2731368,
"is_parallel": true,
"self": 348.2564347824268
}
}
}
}
},
"steps_from_proto": {
"total": 0.039764535004906065,
"count": 48,
"is_parallel": true,
"self": 0.008382477997656679,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.031382057007249387,
"count": 192,
"is_parallel": true,
"self": 0.031382057007249387
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 2153.210631092451,
"count": 341422,
"self": 47.79449909016421,
"children": {
"process_trajectory": {
"total": 1152.3439852402807,
"count": 341422,
"self": 1149.5954861252776,
"children": {
"RLTrainer._checkpoint": {
"total": 2.748499115003142,
"count": 10,
"self": 2.748499115003142
}
}
},
"_update_policy": {
"total": 953.072146762006,
"count": 239,
"self": 478.42174236902156,
"children": {
"TorchPOCAOptimizer.update": {
"total": 474.65040439298446,
"count": 7188,
"self": 474.65040439298446
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.800008461344987e-07,
"count": 1,
"self": 7.800008461344987e-07
},
"TrainerController._save_models": {
"total": 0.2943958619998739,
"count": 1,
"self": 0.006281059999309946,
"children": {
"RLTrainer._checkpoint": {
"total": 0.28811480200056394,
"count": 1,
"self": 0.28811480200056394
}
}
}
}
}
}
}