poca-SoccerTwos / run_logs /timers.json
CoreyMorris's picture
First Push
47a00ff
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.7212694883346558,
"min": 1.647125482559204,
"max": 3.295701742172241,
"count": 2588
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 34535.55078125,
"min": 17049.26953125,
"max": 140515.859375,
"count": 2588
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 58.79761904761905,
"min": 40.31404958677686,
"max": 999.0,
"count": 2588
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19756.0,
"min": 14060.0,
"max": 28132.0,
"count": 2588
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1661.3725387794098,
"min": 1192.1137038304564,
"max": 1694.7295408276657,
"count": 1788
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 279110.58651494083,
"min": 2384.227407660913,
"max": 384134.8325698135,
"count": 1788
},
"SoccerTwos.Step.mean": {
"value": 25879980.0,
"min": 9084.0,
"max": 25879980.0,
"count": 2588
},
"SoccerTwos.Step.sum": {
"value": 25879980.0,
"min": 9084.0,
"max": 25879980.0,
"count": 2588
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.02499029040336609,
"min": -0.14508381485939026,
"max": 0.1613924354314804,
"count": 2588
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -4.223359107971191,
"min": -23.503578186035156,
"max": 21.599212646484375,
"count": 2588
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.02887924388051033,
"min": -0.14518895745277405,
"max": 0.15483121573925018,
"count": 2588
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -4.880592346191406,
"min": -23.520610809326172,
"max": 21.369853973388672,
"count": 2588
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2588
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2588
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.10027692233316997,
"min": -0.6125411776935353,
"max": 0.4631360054016113,
"count": 2588
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 16.946799874305725,
"min": -68.4883998632431,
"max": 58.71160018444061,
"count": 2588
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.10027692233316997,
"min": -0.6125411776935353,
"max": 0.4631360054016113,
"count": 2588
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 16.946799874305725,
"min": -68.4883998632431,
"max": 58.71160018444061,
"count": 2588
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2588
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2588
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.01471707941382192,
"min": 0.010004482062261863,
"max": 0.026951852050842717,
"count": 1224
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.01471707941382192,
"min": 0.010004482062261863,
"max": 0.026951852050842717,
"count": 1224
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.09931258608897527,
"min": 3.1994166667844346e-10,
"max": 0.1213124692440033,
"count": 1224
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.09931258608897527,
"min": 3.1994166667844346e-10,
"max": 0.1213124692440033,
"count": 1224
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.10058390721678734,
"min": 3.4912994069843016e-10,
"max": 0.12339624067147573,
"count": 1224
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.10058390721678734,
"min": 3.4912994069843016e-10,
"max": 0.12339624067147573,
"count": 1224
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 1224
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 1224
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 1224
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 1224
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 1224
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 1224
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1675650130",
"python_version": "3.9.16 (main, Jan 11 2023, 16:05:54) \n[GCC 11.2.0]",
"command_line_arguments": "/home/corey/anaconda3/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos --run-id=SoccerTwosa --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1675708236"
},
"total": 58105.543305226005,
"count": 1,
"self": 0.09773968300578417,
"children": {
"run_training.setup": {
"total": 0.008346703999961846,
"count": 1,
"self": 0.008346703999961846
},
"TrainerController.start_learning": {
"total": 58105.437218839,
"count": 1,
"self": 39.78711152682081,
"children": {
"TrainerController._reset_env": {
"total": 12.791768706002813,
"count": 130,
"self": 12.791768706002813
},
"TrainerController.advance": {
"total": 58052.63279462217,
"count": 1729821,
"self": 38.72335009535891,
"children": {
"env_step": {
"total": 46961.93250139916,
"count": 1729821,
"self": 39455.99591052753,
"children": {
"SubprocessEnvManager._take_step": {
"total": 7478.727212097932,
"count": 1729821,
"self": 208.8250869094045,
"children": {
"TorchPolicy.evaluate": {
"total": 7269.902125188528,
"count": 3302731,
"self": 7269.902125188528
}
}
},
"workers": {
"total": 27.209378773700962,
"count": 1729820,
"self": 0.0,
"children": {
"worker_root": {
"total": 58024.974306515134,
"count": 1729820,
"is_parallel": true,
"self": 23955.69436430854,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.00627252299995007,
"count": 2,
"is_parallel": true,
"self": 0.001897253999914028,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.004375269000036042,
"count": 8,
"is_parallel": true,
"self": 0.004375269000036042
}
}
},
"UnityEnvironment.step": {
"total": 0.038105620000010276,
"count": 1,
"is_parallel": true,
"self": 0.0011472850000586732,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0011037709999754952,
"count": 1,
"is_parallel": true,
"self": 0.0011037709999754952
},
"communicator.exchange": {
"total": 0.03292682699998295,
"count": 1,
"is_parallel": true,
"self": 0.03292682699998295
},
"steps_from_proto": {
"total": 0.002927736999993158,
"count": 2,
"is_parallel": true,
"self": 0.0005652000000395674,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0023625369999535906,
"count": 8,
"is_parallel": true,
"self": 0.0023625369999535906
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 34068.74425991656,
"count": 1729819,
"is_parallel": true,
"self": 2665.6619128927887,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 1329.3729060250018,
"count": 1729819,
"is_parallel": true,
"self": 1329.3729060250018
},
"communicator.exchange": {
"total": 23279.001527161403,
"count": 1729819,
"is_parallel": true,
"self": 23279.001527161403
},
"steps_from_proto": {
"total": 6794.7079138373665,
"count": 3459638,
"is_parallel": true,
"self": 1176.1299878110349,
"children": {
"_process_rank_one_or_two_observation": {
"total": 5618.577926026332,
"count": 13838552,
"is_parallel": true,
"self": 5618.577926026332
}
}
}
}
},
"steps_from_proto": {
"total": 0.5356822900325824,
"count": 258,
"is_parallel": true,
"self": 0.09209280297477562,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.44358948705780676,
"count": 1032,
"is_parallel": true,
"self": 0.44358948705780676
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 11051.976943127655,
"count": 1729820,
"self": 313.1084055634492,
"children": {
"process_trajectory": {
"total": 3258.2094100432396,
"count": 1729820,
"self": 3245.914919233252,
"children": {
"RLTrainer._checkpoint": {
"total": 12.294490809987565,
"count": 51,
"self": 12.294490809987565
}
}
},
"_update_policy": {
"total": 7480.659127520966,
"count": 1224,
"self": 4175.078776194107,
"children": {
"TorchPOCAOptimizer.update": {
"total": 3305.5803513268584,
"count": 36720,
"self": 3305.5803513268584
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.020004654303193e-07,
"count": 1,
"self": 9.020004654303193e-07
},
"TrainerController._save_models": {
"total": 0.2255430820005131,
"count": 1,
"self": 0.0012028569981339388,
"children": {
"RLTrainer._checkpoint": {
"total": 0.22434022500237916,
"count": 1,
"self": 0.22434022500237916
}
}
}
}
}
}
}