ppo-Huggy / run_logs /timers.json
eikoenchine's picture
Huggy
0a5a951
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4071847200393677,
"min": 1.4071847200393677,
"max": 1.42912757396698,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70436.6328125,
"min": 68770.1328125,
"max": 77197.5703125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 109.20264317180617,
"min": 97.90693069306931,
"max": 397.3015873015873,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49578.0,
"min": 48964.0,
"max": 50193.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999933.0,
"min": 49794.0,
"max": 1999933.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999933.0,
"min": 49794.0,
"max": 1999933.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.3077268600463867,
"min": -0.030434073880314827,
"max": 2.407374143600464,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1047.7080078125,
"min": -3.8042593002319336,
"max": 1177.57958984375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.490399238559118,
"min": 1.6506823596954345,
"max": 3.802232749136534,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1584.6412543058395,
"min": 206.33529496192932,
"max": 1823.6968754529953,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.490399238559118,
"min": 1.6506823596954345,
"max": 3.802232749136534,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1584.6412543058395,
"min": 206.33529496192932,
"max": 1823.6968754529953,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.015747409302912276,
"min": 0.013982561259520784,
"max": 0.021250190981663763,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.03149481860582455,
"min": 0.02796512251904157,
"max": 0.059090837721790496,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.05063373458882173,
"min": 0.022633647587564257,
"max": 0.05494891125708819,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.10126746917764345,
"min": 0.046650676801800725,
"max": 0.15820129662752153,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 4.427573524174992e-06,
"min": 4.427573524174992e-06,
"max": 0.00029536665154445004,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 8.855147048349984e-06,
"min": 8.855147048349984e-06,
"max": 0.0008439960186679999,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10147582500000002,
"min": 0.10147582500000002,
"max": 0.19845555000000006,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20295165000000004,
"min": 0.20295165000000004,
"max": 0.5813320000000002,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 8.364366749999985e-05,
"min": 8.364366749999985e-05,
"max": 0.004922931945,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0001672873349999997,
"min": 0.0001672873349999997,
"max": 0.014068466800000002,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1690330298",
"python_version": "3.10.6 (main, May 29 2023, 11:10:38) [GCC 11.3.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics --resume",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1690332245"
},
"total": 1946.6513342400003,
"count": 1,
"self": 0.32121696100057306,
"children": {
"run_training.setup": {
"total": 0.03150228100003005,
"count": 1,
"self": 0.03150228100003005
},
"TrainerController.start_learning": {
"total": 1946.2986149979997,
"count": 1,
"self": 4.127312205040198,
"children": {
"TrainerController._reset_env": {
"total": 3.8991622580000467,
"count": 1,
"self": 3.8991622580000467
},
"TrainerController.advance": {
"total": 1938.1598381889598,
"count": 231126,
"self": 3.9426333450844595,
"children": {
"env_step": {
"total": 1465.8584584429796,
"count": 231126,
"self": 1201.001441823927,
"children": {
"SubprocessEnvManager._take_step": {
"total": 262.1855099190319,
"count": 231126,
"self": 15.074062901122602,
"children": {
"TorchPolicy.evaluate": {
"total": 247.1114470179093,
"count": 223001,
"self": 247.1114470179093
}
}
},
"workers": {
"total": 2.6715067000204726,
"count": 231126,
"self": 0.0,
"children": {
"worker_root": {
"total": 1938.926437993949,
"count": 231126,
"is_parallel": true,
"self": 971.8683746948302,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0008233069999050713,
"count": 1,
"is_parallel": true,
"self": 0.00022934999992685334,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005939569999782179,
"count": 2,
"is_parallel": true,
"self": 0.0005939569999782179
}
}
},
"UnityEnvironment.step": {
"total": 0.031867083999941315,
"count": 1,
"is_parallel": true,
"self": 0.00016392200006976054,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00017166299994642031,
"count": 1,
"is_parallel": true,
"self": 0.00017166299994642031
},
"communicator.exchange": {
"total": 0.03110883699991973,
"count": 1,
"is_parallel": true,
"self": 0.03110883699991973
},
"steps_from_proto": {
"total": 0.00042266200000540266,
"count": 1,
"is_parallel": true,
"self": 0.00012899300008939463,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00029366899991600803,
"count": 2,
"is_parallel": true,
"self": 0.00029366899991600803
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 967.0580632991189,
"count": 231125,
"is_parallel": true,
"self": 26.94094800516416,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 49.70449671704239,
"count": 231125,
"is_parallel": true,
"self": 49.70449671704239
},
"communicator.exchange": {
"total": 825.6199269889598,
"count": 231125,
"is_parallel": true,
"self": 825.6199269889598
},
"steps_from_proto": {
"total": 64.7926915879525,
"count": 231125,
"is_parallel": true,
"self": 25.46672496991323,
"children": {
"_process_rank_one_or_two_observation": {
"total": 39.325966618039274,
"count": 462250,
"is_parallel": true,
"self": 39.325966618039274
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 468.3587464008958,
"count": 231126,
"self": 5.710515309867333,
"children": {
"process_trajectory": {
"total": 116.20865000202889,
"count": 231126,
"self": 115.01167593402829,
"children": {
"RLTrainer._checkpoint": {
"total": 1.1969740680005998,
"count": 10,
"self": 1.1969740680005998
}
}
},
"_update_policy": {
"total": 346.4395810889996,
"count": 96,
"self": 291.48218570800566,
"children": {
"TorchPPOOptimizer.update": {
"total": 54.957395380993944,
"count": 2880,
"self": 54.957395380993944
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.769996722752694e-07,
"count": 1,
"self": 9.769996722752694e-07
},
"TrainerController._save_models": {
"total": 0.11230136899985155,
"count": 1,
"self": 0.0023018029996819678,
"children": {
"RLTrainer._checkpoint": {
"total": 0.10999956600016958,
"count": 1,
"self": 0.10999956600016958
}
}
}
}
}
}
}