Pyramids / run_logs /timers.json
hannahh7's picture
First try
0cabc25
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.31846192479133606,
"min": 0.31712907552719116,
"max": 1.380419373512268,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 9528.380859375,
"min": 9528.380859375,
"max": 41876.40234375,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989928.0,
"min": 29952.0,
"max": 989928.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989928.0,
"min": 29952.0,
"max": 989928.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.5974531769752502,
"min": -0.1484963595867157,
"max": 0.5992048382759094,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 166.68943786621094,
"min": -35.78762435913086,
"max": 168.37655639648438,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.01993049681186676,
"min": 0.003600776195526123,
"max": 0.5776336193084717,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 5.560608863830566,
"min": 0.936201810836792,
"max": 136.899169921875,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.07101942357465824,
"min": 0.06457643765800943,
"max": 0.07451330850757995,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.9942719300452153,
"min": 0.5215931595530596,
"max": 1.0263331604849857,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.01761896005302246,
"min": 0.0007128264535664507,
"max": 0.01771979556584937,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.24666544074231447,
"min": 0.009266743896363859,
"max": 0.24807713792189118,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.331540413328571e-06,
"min": 7.331540413328571e-06,
"max": 0.00029515063018788575,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.0001026415657866,
"min": 0.0001026415657866,
"max": 0.0035079959306680993,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10244381428571428,
"min": 0.10244381428571428,
"max": 0.19838354285714285,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.4342134,
"min": 1.3886848,
"max": 2.569331900000001,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.00025413704714285715,
"min": 0.00025413704714285715,
"max": 0.00983851593142857,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.00355791866,
"min": 0.00355791866,
"max": 0.11695625680999998,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.016155779361724854,
"min": 0.016155779361724854,
"max": 0.6117459535598755,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.22618091106414795,
"min": 0.22618091106414795,
"max": 4.282221794128418,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 324.1530612244898,
"min": 306.96875,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 31767.0,
"min": 15984.0,
"max": 33079.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.655430595333479,
"min": -1.0000000521540642,
"max": 1.655430595333479,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 162.23219834268093,
"min": -30.730001732707024,
"max": 162.23219834268093,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.655430595333479,
"min": -1.0000000521540642,
"max": 1.655430595333479,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 162.23219834268093,
"min": -30.730001732707024,
"max": 162.23219834268093,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.05385339676404885,
"min": 0.05242809605503377,
"max": 12.16811584867537,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 5.277632882876787,
"min": 4.855591664090753,
"max": 194.68985357880592,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1686556903",
"python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1686559229"
},
"total": 2325.9307431629995,
"count": 1,
"self": 0.43891532299949176,
"children": {
"run_training.setup": {
"total": 0.04168815100001666,
"count": 1,
"self": 0.04168815100001666
},
"TrainerController.start_learning": {
"total": 2325.450139689,
"count": 1,
"self": 1.6588341330548246,
"children": {
"TrainerController._reset_env": {
"total": 4.208752058999835,
"count": 1,
"self": 4.208752058999835
},
"TrainerController.advance": {
"total": 2319.487763813945,
"count": 63853,
"self": 1.6362681219602564,
"children": {
"env_step": {
"total": 1657.7779834429743,
"count": 63853,
"self": 1531.9953808459175,
"children": {
"SubprocessEnvManager._take_step": {
"total": 124.7582864710239,
"count": 63853,
"self": 5.141384641045306,
"children": {
"TorchPolicy.evaluate": {
"total": 119.61690182997859,
"count": 62556,
"self": 119.61690182997859
}
}
},
"workers": {
"total": 1.0243161260329998,
"count": 63853,
"self": 0.0,
"children": {
"worker_root": {
"total": 2319.581137104038,
"count": 63853,
"is_parallel": true,
"self": 913.2930507030565,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0019038820000787382,
"count": 1,
"is_parallel": true,
"self": 0.0005755860001954716,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013282959998832666,
"count": 8,
"is_parallel": true,
"self": 0.0013282959998832666
}
}
},
"UnityEnvironment.step": {
"total": 0.05056612099997437,
"count": 1,
"is_parallel": true,
"self": 0.0005523979998542927,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0004768680000779568,
"count": 1,
"is_parallel": true,
"self": 0.0004768680000779568
},
"communicator.exchange": {
"total": 0.047670763000041916,
"count": 1,
"is_parallel": true,
"self": 0.047670763000041916
},
"steps_from_proto": {
"total": 0.001866092000000208,
"count": 1,
"is_parallel": true,
"self": 0.0003802499995799735,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014858420004202344,
"count": 8,
"is_parallel": true,
"self": 0.0014858420004202344
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1406.2880864009817,
"count": 63852,
"is_parallel": true,
"self": 34.90020701278331,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 23.164081639086817,
"count": 63852,
"is_parallel": true,
"self": 23.164081639086817
},
"communicator.exchange": {
"total": 1241.7657896400524,
"count": 63852,
"is_parallel": true,
"self": 1241.7657896400524
},
"steps_from_proto": {
"total": 106.45800810905916,
"count": 63852,
"is_parallel": true,
"self": 21.50874675705336,
"children": {
"_process_rank_one_or_two_observation": {
"total": 84.9492613520058,
"count": 510816,
"is_parallel": true,
"self": 84.9492613520058
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 660.0735122490105,
"count": 63853,
"self": 3.0032061210431493,
"children": {
"process_trajectory": {
"total": 113.14352601996484,
"count": 63853,
"self": 112.93943343296428,
"children": {
"RLTrainer._checkpoint": {
"total": 0.20409258700055943,
"count": 2,
"self": 0.20409258700055943
}
}
},
"_update_policy": {
"total": 543.9267801080025,
"count": 450,
"self": 346.762305897053,
"children": {
"TorchPPOOptimizer.update": {
"total": 197.1644742109495,
"count": 22845,
"self": 197.1644742109495
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.0739995559561066e-06,
"count": 1,
"self": 1.0739995559561066e-06
},
"TrainerController._save_models": {
"total": 0.09478860900071595,
"count": 1,
"self": 0.0013730960008615511,
"children": {
"RLTrainer._checkpoint": {
"total": 0.0934155129998544,
"count": 1,
"self": 0.0934155129998544
}
}
}
}
}
}
}