Pyramids-PPO / run_logs /timers.json
dbaibak's picture
Adjusted parameters
0bc779f
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.172193706035614,
"min": 0.1647656112909317,
"max": 1.4083733558654785,
"count": 100
},
"Pyramids.Policy.Entropy.sum": {
"value": 5141.01513671875,
"min": 4979.8759765625,
"max": 42724.4140625,
"count": 100
},
"Pyramids.Step.mean": {
"value": 2999981.0,
"min": 29952.0,
"max": 2999981.0,
"count": 100
},
"Pyramids.Step.sum": {
"value": 2999981.0,
"min": 29952.0,
"max": 2999981.0,
"count": 100
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.6161277294158936,
"min": -0.09271835535764694,
"max": 0.6173363924026489,
"count": 100
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 174.36415100097656,
"min": -22.530559539794922,
"max": 174.36415100097656,
"count": 100
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": -0.014366395771503448,
"min": -1.7395180463790894,
"max": 1.7319082021713257,
"count": 100
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": -4.065690040588379,
"min": -455.75372314453125,
"max": 453.75994873046875,
"count": 100
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.04603456548027074,
"min": 0.04314542232702175,
"max": 0.0513265693273956,
"count": 100
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.6905184822040611,
"min": 0.33428702012669026,
"max": 0.7394997566007078,
"count": 100
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.015738900086532035,
"min": 0.00041164355064317423,
"max": 0.40088433486276437,
"count": 100
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.23608350129798053,
"min": 0.004528079057074917,
"max": 5.6123806880787015,
"count": 100
},
"Pyramids.Policy.LearningRate.mean": {
"value": 4.911817311399989e-07,
"min": 4.911817311399989e-07,
"max": 9.946118149120001e-05,
"count": 100
},
"Pyramids.Policy.LearningRate.sum": {
"value": 7.367725967099983e-06,
"min": 7.367725967099983e-06,
"max": 0.0013274454725547,
"count": 100
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10049108222222224,
"min": 0.10049108222222224,
"max": 0.19946118095238097,
"count": 100
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.5073662333333335,
"min": 1.3897045333333333,
"max": 2.8274453000000004,
"count": 100
},
"Pyramids.Policy.Beta.mean": {
"value": 5.905911399999989e-05,
"min": 5.905911399999989e-05,
"max": 0.009946171977142856,
"count": 100
},
"Pyramids.Policy.Beta.sum": {
"value": 0.0008858867099999984,
"min": 0.0008858867099999984,
"max": 0.13276178547,
"count": 100
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.026145173236727715,
"min": 0.026145173236727715,
"max": 0.5980517268180847,
"count": 100
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.39217761158943176,
"min": 0.36809539794921875,
"max": 4.186362266540527,
"count": 100
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 310.4,
"min": 289.4848484848485,
"max": 999.0,
"count": 100
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 29488.0,
"min": 15984.0,
"max": 34277.0,
"count": 100
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.6895999808060496,
"min": -1.0000000521540642,
"max": 1.6911179810762405,
"count": 100
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 160.5119981765747,
"min": -32.000001668930054,
"max": 169.11179810762405,
"count": 100
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.6895999808060496,
"min": -1.0000000521540642,
"max": 1.6911179810762405,
"count": 100
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 160.5119981765747,
"min": -32.000001668930054,
"max": 169.11179810762405,
"count": 100
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.08286011184835317,
"min": 0.07857821287703701,
"max": 11.01363092660904,
"count": 100
},
"Pyramids.Policy.RndReward.sum": {
"value": 7.8717106255935505,
"min": 7.174652462999802,
"max": 176.21809482574463,
"count": 100
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1673783626",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics --force",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1673791034"
},
"total": 7408.060608971001,
"count": 1,
"self": 0.4775461550007094,
"children": {
"run_training.setup": {
"total": 0.11163453599965578,
"count": 1,
"self": 0.11163453599965578
},
"TrainerController.start_learning": {
"total": 7407.4714282800005,
"count": 1,
"self": 5.1130147120175025,
"children": {
"TrainerController._reset_env": {
"total": 6.665063098000246,
"count": 1,
"self": 6.665063098000246
},
"TrainerController.advance": {
"total": 7395.578573279985,
"count": 192507,
"self": 5.180274127194934,
"children": {
"env_step": {
"total": 4772.939104372194,
"count": 192507,
"self": 4390.172587254402,
"children": {
"SubprocessEnvManager._take_step": {
"total": 379.5830282929637,
"count": 192507,
"self": 15.347926996098977,
"children": {
"TorchPolicy.evaluate": {
"total": 364.23510129686474,
"count": 187549,
"self": 121.69350056174335,
"children": {
"TorchPolicy.sample_actions": {
"total": 242.5416007351214,
"count": 187549,
"self": 242.5416007351214
}
}
}
}
},
"workers": {
"total": 3.1834888248286006,
"count": 192507,
"self": 0.0,
"children": {
"worker_root": {
"total": 7391.3437224265335,
"count": 192507,
"is_parallel": true,
"self": 3360.819595076673,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0019106769996142248,
"count": 1,
"is_parallel": true,
"self": 0.0006517159990835353,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0012589610005306895,
"count": 8,
"is_parallel": true,
"self": 0.0012589610005306895
}
}
},
"UnityEnvironment.step": {
"total": 0.04981924700041418,
"count": 1,
"is_parallel": true,
"self": 0.0005316970000421861,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0004576800001814263,
"count": 1,
"is_parallel": true,
"self": 0.0004576800001814263
},
"communicator.exchange": {
"total": 0.04700930600029096,
"count": 1,
"is_parallel": true,
"self": 0.04700930600029096
},
"steps_from_proto": {
"total": 0.0018205639998996048,
"count": 1,
"is_parallel": true,
"self": 0.0004515860009632888,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001368977998936316,
"count": 8,
"is_parallel": true,
"self": 0.001368977998936316
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 4030.5241273498605,
"count": 192506,
"is_parallel": true,
"self": 93.57510296254623,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 77.95089503732379,
"count": 192506,
"is_parallel": true,
"self": 77.95089503732379
},
"communicator.exchange": {
"total": 3511.6009194647468,
"count": 192506,
"is_parallel": true,
"self": 3511.6009194647468
},
"steps_from_proto": {
"total": 347.3972098852437,
"count": 192506,
"is_parallel": true,
"self": 78.90584047933771,
"children": {
"_process_rank_one_or_two_observation": {
"total": 268.49136940590597,
"count": 1540048,
"is_parallel": true,
"self": 268.49136940590597
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 2617.4591947805957,
"count": 192507,
"self": 9.635674524552996,
"children": {
"process_trajectory": {
"total": 479.09892203501477,
"count": 192507,
"self": 478.4984932260122,
"children": {
"RLTrainer._checkpoint": {
"total": 0.6004288090025511,
"count": 6,
"self": 0.6004288090025511
}
}
},
"_update_policy": {
"total": 2128.724598221028,
"count": 1391,
"self": 996.1768557010346,
"children": {
"TorchPPOOptimizer.update": {
"total": 1132.5477425199933,
"count": 56690,
"self": 1132.5477425199933
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.919986041495577e-07,
"count": 1,
"self": 9.919986041495577e-07
},
"TrainerController._save_models": {
"total": 0.11477619799916283,
"count": 1,
"self": 0.0014028049990884028,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11337339300007443,
"count": 1,
"self": 0.11337339300007443
}
}
}
}
}
}
}