SoccerTwos / run_logs /timers.json
eduiqe's picture
First Push
4b7b78e
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.20078444480896,
"min": 3.185899257659912,
"max": 3.295761823654175,
"count": 50
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 72824.25,
"min": 35182.734375,
"max": 117208.53125,
"count": 50
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 900.4,
"min": 564.3333333333334,
"max": 999.0,
"count": 50
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 18008.0,
"min": 7992.0,
"max": 28228.0,
"count": 50
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1201.9925436475523,
"min": 1197.1270564742363,
"max": 1201.9925436475523,
"count": 41
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 2403.9850872951047,
"min": 2395.4673115163596,
"max": 9614.882518281993,
"count": 41
},
"SoccerTwos.Step.mean": {
"value": 499750.0,
"min": 9746.0,
"max": 499750.0,
"count": 50
},
"SoccerTwos.Step.sum": {
"value": 499750.0,
"min": 9746.0,
"max": 499750.0,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.007537536323070526,
"min": -0.0092100128531456,
"max": 0.09207545220851898,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -0.08291289955377579,
"min": -0.11052015423774719,
"max": 1.3811317682266235,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.007302678655833006,
"min": -0.009990900754928589,
"max": 0.09211859852075577,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -0.08032946288585663,
"min": -0.11989080905914307,
"max": 1.3817789554595947,
"count": 50
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 50
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.18181818181818182,
"min": -0.3333333333333333,
"max": 0.1885999987522761,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -2.0,
"min": -4.0,
"max": 2.470800042152405,
"count": 50
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.18181818181818182,
"min": -0.3333333333333333,
"max": 0.1885999987522761,
"count": 50
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -2.0,
"min": -4.0,
"max": 2.470800042152405,
"count": 50
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 50
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 50
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.02015251002740115,
"min": 0.011811718170065433,
"max": 0.020813787650937835,
"count": 23
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.02015251002740115,
"min": 0.011811718170065433,
"max": 0.020813787650937835,
"count": 23
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.0012056249504288038,
"min": 6.8407710447597e-05,
"max": 0.0033159471000544726,
"count": 23
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.0012056249504288038,
"min": 6.8407710447597e-05,
"max": 0.0033159471000544726,
"count": 23
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.0012118047801777721,
"min": 6.656067380390595e-05,
"max": 0.0033089817967265843,
"count": 23
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.0012118047801777721,
"min": 6.656067380390595e-05,
"max": 0.0033089817967265843,
"count": 23
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 23
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 23
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 23
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 23
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 23
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 23
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1680730630",
"python_version": "3.9.16 | packaged by conda-forge | (main, Feb 1 2023, 21:39:03) \n[GCC 11.3.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn /content/ml-agents/config/poca/SoccerTwos.yaml --env /content/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1680737944"
},
"total": 7313.7636508040005,
"count": 1,
"self": 0.4370294999998805,
"children": {
"run_training.setup": {
"total": 0.03536977800013119,
"count": 1,
"self": 0.03536977800013119
},
"TrainerController.start_learning": {
"total": 7313.291251526,
"count": 1,
"self": 1.0324574249725629,
"children": {
"TrainerController._reset_env": {
"total": 4.164520597000092,
"count": 3,
"self": 4.164520597000092
},
"TrainerController.advance": {
"total": 7307.841693728027,
"count": 32940,
"self": 1.151269343156855,
"children": {
"env_step": {
"total": 7078.280401643975,
"count": 32940,
"self": 6866.089117045961,
"children": {
"SubprocessEnvManager._take_step": {
"total": 211.53437018000022,
"count": 32940,
"self": 6.015100069978416,
"children": {
"TorchPolicy.evaluate": {
"total": 205.5192701100218,
"count": 65528,
"self": 205.5192701100218
}
}
},
"workers": {
"total": 0.6569144180141393,
"count": 32940,
"self": 0.0,
"children": {
"worker_root": {
"total": 7299.216771676062,
"count": 32940,
"is_parallel": true,
"self": 577.6291194390124,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.003633663000073284,
"count": 2,
"is_parallel": true,
"self": 0.0007954440002322372,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0028382189998410468,
"count": 8,
"is_parallel": true,
"self": 0.0028382189998410468
}
}
},
"UnityEnvironment.step": {
"total": 0.16920685600007346,
"count": 1,
"is_parallel": true,
"self": 0.00039719300002616365,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.003184289999808243,
"count": 1,
"is_parallel": true,
"self": 0.003184289999808243
},
"communicator.exchange": {
"total": 0.16003474700005427,
"count": 1,
"is_parallel": true,
"self": 0.16003474700005427
},
"steps_from_proto": {
"total": 0.005590626000184784,
"count": 2,
"is_parallel": true,
"self": 0.0008009350001429993,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.004789691000041785,
"count": 8,
"is_parallel": true,
"self": 0.004789691000041785
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 6721.58189364505,
"count": 32939,
"is_parallel": true,
"self": 17.69428333822634,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 97.15385266897397,
"count": 32939,
"is_parallel": true,
"self": 97.15385266897397
},
"communicator.exchange": {
"total": 6351.65877903288,
"count": 32939,
"is_parallel": true,
"self": 6351.65877903288
},
"steps_from_proto": {
"total": 255.0749786049687,
"count": 65878,
"is_parallel": true,
"self": 33.99538617702228,
"children": {
"_process_rank_one_or_two_observation": {
"total": 221.0795924279464,
"count": 263512,
"is_parallel": true,
"self": 221.0795924279464
}
}
}
}
},
"steps_from_proto": {
"total": 0.005758592000347562,
"count": 4,
"is_parallel": true,
"self": 0.0008720690002519405,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.004886523000095622,
"count": 16,
"is_parallel": true,
"self": 0.004886523000095622
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 228.41002274089533,
"count": 32940,
"self": 7.72681306698928,
"children": {
"process_trajectory": {
"total": 56.219104672907406,
"count": 32940,
"self": 55.8566886259066,
"children": {
"RLTrainer._checkpoint": {
"total": 0.36241604700080643,
"count": 1,
"self": 0.36241604700080643
}
}
},
"_update_policy": {
"total": 164.46410500099864,
"count": 23,
"self": 105.91964750200304,
"children": {
"TorchPOCAOptimizer.update": {
"total": 58.5444574989956,
"count": 699,
"self": 58.5444574989956
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.380000847158954e-07,
"count": 1,
"self": 9.380000847158954e-07
},
"TrainerController._save_models": {
"total": 0.2525788380007725,
"count": 1,
"self": 0.0017644560011831345,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2508143819995894,
"count": 1,
"self": 0.2508143819995894
}
}
}
}
}
}
}