ppo-Huggy / run_logs /timers.json
bguan's picture
Huggy
25f9f1a
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4023890495300293,
"min": 1.4023890495300293,
"max": 1.4281084537506104,
"count": 37
},
"Huggy.Policy.Entropy.sum": {
"value": 70334.015625,
"min": 66132.640625,
"max": 72420.46875,
"count": 37
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 104.39662447257383,
"min": 83.65423728813559,
"max": 236.85245901639345,
"count": 37
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49484.0,
"min": 43344.0,
"max": 50136.0,
"count": 37
},
"Huggy.Step.mean": {
"value": 1999944.0,
"min": 199676.0,
"max": 1999944.0,
"count": 37
},
"Huggy.Step.sum": {
"value": 1999944.0,
"min": 199676.0,
"max": 1999944.0,
"count": 37
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.379626989364624,
"min": 0.7843582034111023,
"max": 2.4256348609924316,
"count": 37
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1127.9432373046875,
"min": 142.75318908691406,
"max": 1414.536376953125,
"count": 37
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.650795543495613,
"min": 3.2260666502701056,
"max": 3.877175527103877,
"count": 37
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1730.4770876169205,
"min": 587.1441303491592,
"max": 2287.5335609912872,
"count": 37
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.650795543495613,
"min": 3.2260666502701056,
"max": 3.877175527103877,
"count": 37
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1730.4770876169205,
"min": 587.1441303491592,
"max": 2287.5335609912872,
"count": 37
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.014996173076603251,
"min": 0.013167249064523882,
"max": 0.020709680289534543,
"count": 37
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.029992346153206503,
"min": 0.026334498129047763,
"max": 0.05911566048744135,
"count": 37
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.045575076589981715,
"min": 0.025805531131724516,
"max": 0.057734017136196296,
"count": 37
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.09115015317996343,
"min": 0.05161106226344903,
"max": 0.1605233009904623,
"count": 37
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.3168238944249927e-06,
"min": 3.3168238944249927e-06,
"max": 0.0002718459843846749,
"count": 37
},
"Huggy.Policy.LearningRate.sum": {
"value": 6.633647788849985e-06,
"min": 6.633647788849985e-06,
"max": 0.0007552620482460001,
"count": 37
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10110557500000002,
"min": 0.10110557500000002,
"max": 0.190615325,
"count": 37
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20221115000000003,
"min": 0.20221115000000003,
"max": 0.5517540000000001,
"count": 37
},
"Huggy.Policy.Beta.mean": {
"value": 6.51681924999999e-05,
"min": 6.51681924999999e-05,
"max": 0.0045317047175,
"count": 37
},
"Huggy.Policy.Beta.sum": {
"value": 0.0001303363849999998,
"min": 0.0001303363849999998,
"max": 0.012592524599999997,
"count": 37
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 37
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 37
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1670378092",
"python_version": "3.8.15 (default, Oct 12 2022, 19:14:39) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --resume --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1670380191"
},
"total": 2099.0523550870003,
"count": 1,
"self": 0.706283082000482,
"children": {
"run_training.setup": {
"total": 0.1132985029998963,
"count": 1,
"self": 0.1132985029998963
},
"TrainerController.start_learning": {
"total": 2098.232773502,
"count": 1,
"self": 4.027508129121998,
"children": {
"TrainerController._reset_env": {
"total": 6.496375311000065,
"count": 1,
"self": 6.496375311000065
},
"TrainerController.advance": {
"total": 2087.5186709938775,
"count": 214437,
"self": 3.9786571049266968,
"children": {
"env_step": {
"total": 1640.357965764068,
"count": 214437,
"self": 1370.1144414502533,
"children": {
"SubprocessEnvManager._take_step": {
"total": 267.8257691729359,
"count": 214437,
"self": 13.82410443382173,
"children": {
"TorchPolicy.evaluate": {
"total": 254.00166473911418,
"count": 205618,
"self": 64.41607159819364,
"children": {
"TorchPolicy.sample_actions": {
"total": 189.58559314092054,
"count": 205618,
"self": 189.58559314092054
}
}
}
}
},
"workers": {
"total": 2.417755140878853,
"count": 214437,
"self": 0.0,
"children": {
"worker_root": {
"total": 2090.295313355978,
"count": 214437,
"is_parallel": true,
"self": 961.6800431169763,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0008852040000419947,
"count": 1,
"is_parallel": true,
"self": 0.000279143000170734,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006060609998712607,
"count": 2,
"is_parallel": true,
"self": 0.0006060609998712607
}
}
},
"UnityEnvironment.step": {
"total": 0.03040271899999425,
"count": 1,
"is_parallel": true,
"self": 0.0002744560001701757,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0001930720000018482,
"count": 1,
"is_parallel": true,
"self": 0.0001930720000018482
},
"communicator.exchange": {
"total": 0.02899748799995905,
"count": 1,
"is_parallel": true,
"self": 0.02899748799995905
},
"steps_from_proto": {
"total": 0.000937702999863177,
"count": 1,
"is_parallel": true,
"self": 0.0003997959995558631,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005379070003073139,
"count": 2,
"is_parallel": true,
"self": 0.0005379070003073139
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1128.6152702390018,
"count": 214436,
"is_parallel": true,
"self": 32.99066089910343,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 71.47997315293492,
"count": 214436,
"is_parallel": true,
"self": 71.47997315293492
},
"communicator.exchange": {
"total": 935.987013890006,
"count": 214436,
"is_parallel": true,
"self": 935.987013890006
},
"steps_from_proto": {
"total": 88.15762229695747,
"count": 214436,
"is_parallel": true,
"self": 36.51521687493505,
"children": {
"_process_rank_one_or_two_observation": {
"total": 51.64240542202242,
"count": 428872,
"is_parallel": true,
"self": 51.64240542202242
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 443.18204812488284,
"count": 214437,
"self": 5.7061210459701215,
"children": {
"process_trajectory": {
"total": 143.46216604091364,
"count": 214437,
"self": 142.89157399091323,
"children": {
"RLTrainer._checkpoint": {
"total": 0.570592050000414,
"count": 4,
"self": 0.570592050000414
}
}
},
"_update_policy": {
"total": 294.0137610379991,
"count": 89,
"self": 243.62497279300715,
"children": {
"TorchPPOOptimizer.update": {
"total": 50.38878824499193,
"count": 2670,
"self": 50.38878824499193
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.5760001588205341e-06,
"count": 1,
"self": 1.5760001588205341e-06
},
"TrainerController._save_models": {
"total": 0.19021749200010163,
"count": 1,
"self": 0.005208949999996548,
"children": {
"RLTrainer._checkpoint": {
"total": 0.18500854200010508,
"count": 1,
"self": 0.18500854200010508
}
}
}
}
}
}
}