ppo-Pyramids-1M / run_logs /timers.json
Emperor's picture
Commit
9b901ec
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.30878549814224243,
"min": 0.30878549814224243,
"max": 1.4065443277359009,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 9233.921875,
"min": 9233.921875,
"max": 42668.9296875,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989983.0,
"min": 29952.0,
"max": 989983.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989983.0,
"min": 29952.0,
"max": 989983.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.6260428428649902,
"min": -0.1239318922162056,
"max": 0.7087668180465698,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 178.42221069335938,
"min": -29.743654251098633,
"max": 201.99855041503906,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.00834602490067482,
"min": -0.03647370636463165,
"max": 0.6505448222160339,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 2.37861704826355,
"min": -10.030269622802734,
"max": 154.1791229248047,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.06878873727857318,
"min": 0.0669038255557699,
"max": 0.07362794883090243,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 1.0318310591785977,
"min": 0.47033216640983455,
"max": 1.0918905346576744,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.016499617166270873,
"min": 0.0010142359632326187,
"max": 0.019295246457944473,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.2474942574940631,
"min": 0.007925119198983432,
"max": 0.2474942574940631,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.453217515626665e-06,
"min": 7.453217515626665e-06,
"max": 0.00029515063018788575,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.00011179826273439997,
"min": 0.00011179826273439997,
"max": 0.0035081453306183,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10248437333333335,
"min": 0.10248437333333335,
"max": 0.19838354285714285,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.5372656000000002,
"min": 1.3691136000000002,
"max": 2.5726459000000004,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.00025818889599999996,
"min": 0.00025818889599999996,
"max": 0.00983851593142857,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.0038728334399999995,
"min": 0.0038728334399999995,
"max": 0.11696123183,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.017046082764863968,
"min": 0.017046082764863968,
"max": 0.6730450391769409,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.2556912302970886,
"min": 0.2498343139886856,
"max": 4.711315155029297,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 307.6666666666667,
"min": 279.50467289719626,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 30459.0,
"min": 15984.0,
"max": 33252.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.6317009940893963,
"min": -1.0000000521540642,
"max": 1.6830953085951716,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 161.53839841485023,
"min": -32.000001668930054,
"max": 180.09119801968336,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.6317009940893963,
"min": -1.0000000521540642,
"max": 1.6830953085951716,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 161.53839841485023,
"min": -32.000001668930054,
"max": 180.09119801968336,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.05351920378179233,
"min": 0.05351920378179233,
"max": 13.652445676038042,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 5.298401174397441,
"min": 5.298401174397441,
"max": 218.43913081660867,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1679431910",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training112 --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1679434154"
},
"total": 2243.598298204999,
"count": 1,
"self": 0.4743138329995418,
"children": {
"run_training.setup": {
"total": 0.10329088599974057,
"count": 1,
"self": 0.10329088599974057
},
"TrainerController.start_learning": {
"total": 2243.020693486,
"count": 1,
"self": 1.3411638967854742,
"children": {
"TrainerController._reset_env": {
"total": 6.524614370999188,
"count": 1,
"self": 6.524614370999188
},
"TrainerController.advance": {
"total": 2235.065435024214,
"count": 64185,
"self": 1.5252545270113842,
"children": {
"env_step": {
"total": 1611.549475112185,
"count": 64185,
"self": 1502.954491310299,
"children": {
"SubprocessEnvManager._take_step": {
"total": 107.8041905680684,
"count": 64185,
"self": 4.912926397321826,
"children": {
"TorchPolicy.evaluate": {
"total": 102.89126417074658,
"count": 62567,
"self": 102.89126417074658
}
}
},
"workers": {
"total": 0.7907932338175669,
"count": 64185,
"self": 0.0,
"children": {
"worker_root": {
"total": 2238.1158219349154,
"count": 64185,
"is_parallel": true,
"self": 850.2850226768842,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0019085300000369898,
"count": 1,
"is_parallel": true,
"self": 0.0006519950002257247,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0012565349998112652,
"count": 8,
"is_parallel": true,
"self": 0.0012565349998112652
}
}
},
"UnityEnvironment.step": {
"total": 0.09338815399951272,
"count": 1,
"is_parallel": true,
"self": 0.0005497140000443324,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0004703879994849558,
"count": 1,
"is_parallel": true,
"self": 0.0004703879994849558
},
"communicator.exchange": {
"total": 0.09047946099963156,
"count": 1,
"is_parallel": true,
"self": 0.09047946099963156
},
"steps_from_proto": {
"total": 0.0018885910003518802,
"count": 1,
"is_parallel": true,
"self": 0.0003943149995393469,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014942760008125333,
"count": 8,
"is_parallel": true,
"self": 0.0014942760008125333
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1387.8307992580312,
"count": 64184,
"is_parallel": true,
"self": 31.01560196646551,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 23.621474511350243,
"count": 64184,
"is_parallel": true,
"self": 23.621474511350243
},
"communicator.exchange": {
"total": 1237.753081751991,
"count": 64184,
"is_parallel": true,
"self": 1237.753081751991
},
"steps_from_proto": {
"total": 95.44064102822449,
"count": 64184,
"is_parallel": true,
"self": 20.51449231301467,
"children": {
"_process_rank_one_or_two_observation": {
"total": 74.92614871520982,
"count": 513472,
"is_parallel": true,
"self": 74.92614871520982
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 621.9907053850175,
"count": 64185,
"self": 2.614478388753014,
"children": {
"process_trajectory": {
"total": 121.5869534382764,
"count": 64185,
"self": 121.32011537627659,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2668380619998061,
"count": 2,
"self": 0.2668380619998061
}
}
},
"_update_policy": {
"total": 497.78927355798805,
"count": 451,
"self": 317.8746718512484,
"children": {
"TorchPPOOptimizer.update": {
"total": 179.91460170673963,
"count": 22830,
"self": 179.91460170673963
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.880004654405639e-07,
"count": 1,
"self": 9.880004654405639e-07
},
"TrainerController._save_models": {
"total": 0.08947920600076031,
"count": 1,
"self": 0.001424603000486968,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08805460300027335,
"count": 1,
"self": 0.08805460300027335
}
}
}
}
}
}
}