ppo-Huggy / run_logs /timers.json
Pearson's picture
Huggy
4cb8d33
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4034042358398438,
"min": 1.4034042358398438,
"max": 1.4278993606567383,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70598.25,
"min": 68025.28125,
"max": 77633.15625,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 92.93560606060606,
"min": 90.24309392265194,
"max": 385.65384615384613,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49070.0,
"min": 48997.0,
"max": 50135.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999995.0,
"min": 49909.0,
"max": 1999995.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999995.0,
"min": 49909.0,
"max": 1999995.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4203755855560303,
"min": 0.00720624765381217,
"max": 2.4304492473602295,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1277.958251953125,
"min": 0.9296059608459473,
"max": 1302.9932861328125,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.627167911692099,
"min": 1.8298869538445806,
"max": 3.798286434501972,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1915.1446573734283,
"min": 236.0554170459509,
"max": 2013.091810286045,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.627167911692099,
"min": 1.8298869538445806,
"max": 3.798286434501972,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1915.1446573734283,
"min": 236.0554170459509,
"max": 2013.091810286045,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.014986180779427136,
"min": 0.013361670521165554,
"max": 0.02078944027792507,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.02997236155885427,
"min": 0.02724682067424737,
"max": 0.06236832083377521,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.051399218601485094,
"min": 0.022085205372422933,
"max": 0.06001239884644746,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.10279843720297019,
"min": 0.044170410744845866,
"max": 0.17883240170776843,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 4.4086735304749984e-06,
"min": 4.4086735304749984e-06,
"max": 0.00029531385156205,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 8.817347060949997e-06,
"min": 8.817347060949997e-06,
"max": 0.0008440450686516501,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.101469525,
"min": 0.101469525,
"max": 0.19843794999999997,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20293905,
"min": 0.20293905,
"max": 0.58134835,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 8.332929749999999e-05,
"min": 8.332929749999999e-05,
"max": 0.004922053705,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00016665859499999998,
"min": 0.00016665859499999998,
"max": 0.014069282664999998,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1675531133",
"python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1675533374"
},
"total": 2241.429959977,
"count": 1,
"self": 0.3997484940000504,
"children": {
"run_training.setup": {
"total": 0.1022608550001678,
"count": 1,
"self": 0.1022608550001678
},
"TrainerController.start_learning": {
"total": 2240.927950628,
"count": 1,
"self": 3.859676335930544,
"children": {
"TrainerController._reset_env": {
"total": 10.050530511000034,
"count": 1,
"self": 10.050530511000034
},
"TrainerController.advance": {
"total": 2226.9087173170697,
"count": 231686,
"self": 4.083868138186062,
"children": {
"env_step": {
"total": 1776.43424532296,
"count": 231686,
"self": 1489.8077645670257,
"children": {
"SubprocessEnvManager._take_step": {
"total": 284.0701775338905,
"count": 231686,
"self": 15.1523059108697,
"children": {
"TorchPolicy.evaluate": {
"total": 268.9178716230208,
"count": 223005,
"self": 67.16117020599836,
"children": {
"TorchPolicy.sample_actions": {
"total": 201.75670141702244,
"count": 223005,
"self": 201.75670141702244
}
}
}
}
},
"workers": {
"total": 2.5563032220436526,
"count": 231686,
"self": 0.0,
"children": {
"worker_root": {
"total": 2233.154298422148,
"count": 231686,
"is_parallel": true,
"self": 1006.7185989272339,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0017314940000687784,
"count": 1,
"is_parallel": true,
"self": 0.00034928800005218363,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013822060000165948,
"count": 2,
"is_parallel": true,
"self": 0.0013822060000165948
}
}
},
"UnityEnvironment.step": {
"total": 0.028182504000142217,
"count": 1,
"is_parallel": true,
"self": 0.0002908690000822389,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0001888339998004085,
"count": 1,
"is_parallel": true,
"self": 0.0001888339998004085
},
"communicator.exchange": {
"total": 0.027006963000076212,
"count": 1,
"is_parallel": true,
"self": 0.027006963000076212
},
"steps_from_proto": {
"total": 0.0006958380001833575,
"count": 1,
"is_parallel": true,
"self": 0.00024024100002861815,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00045559700015473936,
"count": 2,
"is_parallel": true,
"self": 0.00045559700015473936
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1226.435699494914,
"count": 231685,
"is_parallel": true,
"self": 34.299367754644436,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 77.76628107394708,
"count": 231685,
"is_parallel": true,
"self": 77.76628107394708
},
"communicator.exchange": {
"total": 1021.6961835811298,
"count": 231685,
"is_parallel": true,
"self": 1021.6961835811298
},
"steps_from_proto": {
"total": 92.67386708519257,
"count": 231685,
"is_parallel": true,
"self": 40.16238961240424,
"children": {
"_process_rank_one_or_two_observation": {
"total": 52.51147747278833,
"count": 463370,
"is_parallel": true,
"self": 52.51147747278833
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 446.3906038559237,
"count": 231686,
"self": 6.573563969883253,
"children": {
"process_trajectory": {
"total": 144.35830985304347,
"count": 231686,
"self": 143.2650338210433,
"children": {
"RLTrainer._checkpoint": {
"total": 1.0932760320001762,
"count": 10,
"self": 1.0932760320001762
}
}
},
"_update_policy": {
"total": 295.458730032997,
"count": 96,
"self": 242.9906067409977,
"children": {
"TorchPPOOptimizer.update": {
"total": 52.46812329199929,
"count": 2880,
"self": 52.46812329199929
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.999997251317836e-07,
"count": 1,
"self": 7.999997251317836e-07
},
"TrainerController._save_models": {
"total": 0.10902566400000069,
"count": 1,
"self": 0.002185938999900827,
"children": {
"RLTrainer._checkpoint": {
"total": 0.10683972500009986,
"count": 1,
"self": 0.10683972500009986
}
}
}
}
}
}
}