poca-SoccerTwos / run_logs /timers.json
toinsson's picture
First Push
adaa211
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.4070029258728027,
"min": 1.364182472229004,
"max": 3.2957570552825928,
"count": 2251
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 27824.890625,
"min": 14507.8642578125,
"max": 105464.2265625,
"count": 2251
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 49.7979797979798,
"min": 40.31147540983606,
"max": 999.0,
"count": 2251
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19720.0,
"min": 13332.0,
"max": 28468.0,
"count": 2251
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1550.0245450575744,
"min": 1190.0487814065316,
"max": 1641.0483088367112,
"count": 2246
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 306904.85992139974,
"min": 2385.9431164520465,
"max": 388701.1363259206,
"count": 2246
},
"SoccerTwos.Step.mean": {
"value": 22509878.0,
"min": 9250.0,
"max": 22509878.0,
"count": 2251
},
"SoccerTwos.Step.sum": {
"value": 22509878.0,
"min": 9250.0,
"max": 22509878.0,
"count": 2251
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.028588376939296722,
"min": -0.14516468346118927,
"max": 0.22160297632217407,
"count": 2251
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 5.66049861907959,
"min": -26.064918518066406,
"max": 33.42010498046875,
"count": 2251
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.02912515588104725,
"min": -0.14482811093330383,
"max": 0.21980521082878113,
"count": 2251
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 5.766780853271484,
"min": -27.210805892944336,
"max": 34.05256652832031,
"count": 2251
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2251
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2251
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.1643575755032626,
"min": -0.6405809518127215,
"max": 0.5064571443058196,
"count": 2251
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 32.542799949645996,
"min": -65.78759974241257,
"max": 66.88839983940125,
"count": 2251
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.1643575755032626,
"min": -0.6405809518127215,
"max": 0.5064571443058196,
"count": 2251
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 32.542799949645996,
"min": -65.78759974241257,
"max": 66.88839983940125,
"count": 2251
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2251
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2251
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.01851427239695719,
"min": 0.010531172272749245,
"max": 0.02492763645859668,
"count": 1091
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.01851427239695719,
"min": 0.010531172272749245,
"max": 0.02492763645859668,
"count": 1091
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.10829064498345058,
"min": 0.0010078916025425618,
"max": 0.1292997161547343,
"count": 1091
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.10829064498345058,
"min": 0.0010078916025425618,
"max": 0.1292997161547343,
"count": 1091
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.11002712522943815,
"min": 0.0010057883948320523,
"max": 0.13142008284727733,
"count": 1091
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.11002712522943815,
"min": 0.0010057883948320523,
"max": 0.13142008284727733,
"count": 1091
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 1091
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 1091
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 1091
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 1091
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 1091
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 1091
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1675332665",
"python_version": "3.8.16 (default, Jan 17 2023, 23:13:24) \n[GCC 11.2.0]",
"command_line_arguments": "/home/antoine/anaconda3/envs/hf_drl_unit7/bin/mlagents-learn ./ml-agents/config/poca/SoccerTwos.yaml --env=./SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos_1 --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1675364819"
},
"total": 32154.188953462,
"count": 1,
"self": 0.1803494279993174,
"children": {
"run_training.setup": {
"total": 0.00640905700129224,
"count": 1,
"self": 0.00640905700129224
},
"TrainerController.start_learning": {
"total": 32154.002194976998,
"count": 1,
"self": 22.474659589188377,
"children": {
"TrainerController._reset_env": {
"total": 4.29831169298086,
"count": 113,
"self": 4.29831169298086
},
"TrainerController.advance": {
"total": 32127.09101790282,
"count": 1562426,
"self": 20.422603639017325,
"children": {
"env_step": {
"total": 24643.542619113483,
"count": 1562426,
"self": 19942.420775126586,
"children": {
"SubprocessEnvManager._take_step": {
"total": 4686.904853418628,
"count": 1562426,
"self": 111.88414057251794,
"children": {
"TorchPolicy.evaluate": {
"total": 4575.02071284611,
"count": 2828472,
"self": 4575.02071284611
}
}
},
"workers": {
"total": 14.216990568267647,
"count": 1562425,
"self": 0.0,
"children": {
"worker_root": {
"total": 32115.610743571782,
"count": 1562425,
"is_parallel": true,
"self": 14869.433992172224,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0017130690011981642,
"count": 2,
"is_parallel": true,
"self": 0.00035320399911142886,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013598650020867353,
"count": 8,
"is_parallel": true,
"self": 0.0013598650020867353
}
}
},
"UnityEnvironment.step": {
"total": 0.020998721000069054,
"count": 1,
"is_parallel": true,
"self": 0.0009554309981467668,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0006709480003337376,
"count": 1,
"is_parallel": true,
"self": 0.0006709480003337376
},
"communicator.exchange": {
"total": 0.016712819000531454,
"count": 1,
"is_parallel": true,
"self": 0.016712819000531454
},
"steps_from_proto": {
"total": 0.002659523001057096,
"count": 2,
"is_parallel": true,
"self": 0.00043291599831718486,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.002226607002739911,
"count": 8,
"is_parallel": true,
"self": 0.002226607002739911
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 17245.90504891357,
"count": 1562424,
"is_parallel": true,
"self": 1198.7301085013478,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 786.7526528414492,
"count": 1562424,
"is_parallel": true,
"self": 786.7526528414492
},
"communicator.exchange": {
"total": 11804.459724711165,
"count": 1562424,
"is_parallel": true,
"self": 11804.459724711165
},
"steps_from_proto": {
"total": 3455.9625628596077,
"count": 3124848,
"is_parallel": true,
"self": 519.3835088980377,
"children": {
"_process_rank_one_or_two_observation": {
"total": 2936.57905396157,
"count": 12499392,
"is_parallel": true,
"self": 2936.57905396157
}
}
}
}
},
"steps_from_proto": {
"total": 0.27170248599031765,
"count": 224,
"is_parallel": true,
"self": 0.03974755697345245,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.2319549290168652,
"count": 896,
"is_parallel": true,
"self": 0.2319549290168652
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 7463.125795150323,
"count": 1562425,
"self": 156.0927741749765,
"children": {
"process_trajectory": {
"total": 3043.756003937433,
"count": 1562425,
"self": 3037.450239336422,
"children": {
"RLTrainer._checkpoint": {
"total": 6.305764601011106,
"count": 45,
"self": 6.305764601011106
}
}
},
"_update_policy": {
"total": 4263.2770170379135,
"count": 1092,
"self": 2639.284905304594,
"children": {
"TorchPOCAOptimizer.update": {
"total": 1623.9921117333197,
"count": 32763,
"self": 1623.9921117333197
}
}
}
}
}
}
},
"trainer_threads": {
"total": 6.400005077011883e-07,
"count": 1,
"self": 6.400005077011883e-07
},
"TrainerController._save_models": {
"total": 0.1382051520049572,
"count": 1,
"self": 0.0012712670068140142,
"children": {
"RLTrainer._checkpoint": {
"total": 0.13693388499814318,
"count": 1,
"self": 0.13693388499814318
}
}
}
}
}
}
}