testpyramidsrnd / run_logs /timers.json
ryanblak's picture
First Pyramids
be45b9c
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.14909592270851135,
"min": 0.142119362950325,
"max": 1.3751806020736694,
"count": 100
},
"Pyramids.Policy.Entropy.sum": {
"value": 4468.1064453125,
"min": 4267.04931640625,
"max": 41717.48046875,
"count": 100
},
"Pyramids.Step.mean": {
"value": 2999985.0,
"min": 29952.0,
"max": 2999985.0,
"count": 100
},
"Pyramids.Step.sum": {
"value": 2999985.0,
"min": 29952.0,
"max": 2999985.0,
"count": 100
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.7961055040359497,
"min": -0.05584091693162918,
"max": 0.9489213824272156,
"count": 100
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 238.03555297851562,
"min": -13.51350212097168,
"max": 299.8591613769531,
"count": 100
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.007581324782222509,
"min": -0.04869057983160019,
"max": 0.5447742938995361,
"count": 100
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 2.2668161392211914,
"min": -13.828125,
"max": 129.11151123046875,
"count": 100
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.06762961348932651,
"min": 0.06311524328781863,
"max": 0.07364776081554604,
"count": 100
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 1.0144442023398976,
"min": 0.49844704683530194,
"max": 1.1001059048964332,
"count": 100
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.015988681381956363,
"min": 0.0011898969587066316,
"max": 0.018554184833559122,
"count": 100
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.23983022072934543,
"min": 0.010709072628359686,
"max": 0.25975858766982773,
"count": 100
},
"Pyramids.Policy.LearningRate.mean": {
"value": 1.473479508873333e-06,
"min": 1.473479508873333e-06,
"max": 0.00029838354339596195,
"count": 100
},
"Pyramids.Policy.LearningRate.sum": {
"value": 2.2102192633099995e-05,
"min": 2.2102192633099995e-05,
"max": 0.004052979249006933,
"count": 100
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10049112666666665,
"min": 0.10049112666666665,
"max": 0.19946118095238097,
"count": 100
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.5073668999999998,
"min": 1.3962282666666668,
"max": 2.827447566666667,
"count": 100
},
"Pyramids.Policy.Beta.mean": {
"value": 5.9063553999999996e-05,
"min": 5.9063553999999996e-05,
"max": 0.009946171977142856,
"count": 100
},
"Pyramids.Policy.Beta.sum": {
"value": 0.0008859533099999999,
"min": 0.0008859533099999999,
"max": 0.13510420736,
"count": 100
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.0069979471154510975,
"min": 0.0069979471154510975,
"max": 0.4304795563220978,
"count": 100
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.10496920347213745,
"min": 0.09808759391307831,
"max": 3.013356924057007,
"count": 100
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 227.1578947368421,
"min": 200.19736842105263,
"max": 999.0,
"count": 100
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 30212.0,
"min": 15984.0,
"max": 33949.0,
"count": 100
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.7427593881922556,
"min": -1.0000000521540642,
"max": 1.7998026256498538,
"count": 100
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 231.78699862957,
"min": -29.47000178694725,
"max": 273.56999909877777,
"count": 100
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.7427593881922556,
"min": -1.0000000521540642,
"max": 1.7998026256498538,
"count": 100
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 231.78699862957,
"min": -29.47000178694725,
"max": 273.56999909877777,
"count": 100
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.01641664565479579,
"min": 0.015668555812630477,
"max": 9.890646627172828,
"count": 100
},
"Pyramids.Policy.RndReward.sum": {
"value": 2.18341387208784,
"min": 2.0968430308421375,
"max": 158.25034603476524,
"count": 100
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1657374594",
"python_version": "3.7.13 (default, Apr 24 2022, 01:04:09) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./trained-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1657381432"
},
"total": 6838.168301733001,
"count": 1,
"self": 0.43616340900098294,
"children": {
"run_training.setup": {
"total": 0.04169457400007559,
"count": 1,
"self": 0.04169457400007559
},
"TrainerController.start_learning": {
"total": 6837.69044375,
"count": 1,
"self": 4.2923572859044725,
"children": {
"TrainerController._reset_env": {
"total": 9.84102506499994,
"count": 1,
"self": 9.84102506499994
},
"TrainerController.advance": {
"total": 6823.463736410095,
"count": 196272,
"self": 4.565868229988155,
"children": {
"env_step": {
"total": 4687.895862876252,
"count": 196272,
"self": 4372.063818228361,
"children": {
"SubprocessEnvManager._take_step": {
"total": 313.50325280693653,
"count": 196272,
"self": 13.913530251997372,
"children": {
"TorchPolicy.evaluate": {
"total": 299.58972255493916,
"count": 187563,
"self": 102.96363168891742,
"children": {
"TorchPolicy.sample_actions": {
"total": 196.62609086602174,
"count": 187563,
"self": 196.62609086602174
}
}
}
}
},
"workers": {
"total": 2.328791840954864,
"count": 196272,
"self": 0.0,
"children": {
"worker_root": {
"total": 6826.345765419902,
"count": 196272,
"is_parallel": true,
"self": 2756.995291435059,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.010451624999973319,
"count": 1,
"is_parallel": true,
"self": 0.004726090999952248,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.005725534000021071,
"count": 8,
"is_parallel": true,
"self": 0.005725534000021071
}
}
},
"UnityEnvironment.step": {
"total": 0.04666723100001491,
"count": 1,
"is_parallel": true,
"self": 0.0005056300000205738,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00043689999995422113,
"count": 1,
"is_parallel": true,
"self": 0.00043689999995422113
},
"communicator.exchange": {
"total": 0.04403368800001317,
"count": 1,
"is_parallel": true,
"self": 0.04403368800001317
},
"steps_from_proto": {
"total": 0.0016910130000269419,
"count": 1,
"is_parallel": true,
"self": 0.0004470860000083121,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0012439270000186298,
"count": 8,
"is_parallel": true,
"self": 0.0012439270000186298
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 4069.350473984843,
"count": 196271,
"is_parallel": true,
"self": 83.9351488638049,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 68.94728767600509,
"count": 196271,
"is_parallel": true,
"self": 68.94728767600509
},
"communicator.exchange": {
"total": 3637.787480832894,
"count": 196271,
"is_parallel": true,
"self": 3637.787480832894
},
"steps_from_proto": {
"total": 278.6805566121395,
"count": 196271,
"is_parallel": true,
"self": 69.70327419935177,
"children": {
"_process_rank_one_or_two_observation": {
"total": 208.97728241278776,
"count": 1570168,
"is_parallel": true,
"self": 208.97728241278776
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 2131.0020053038547,
"count": 196272,
"self": 8.439120837779683,
"children": {
"process_trajectory": {
"total": 497.3240388450703,
"count": 196272,
"self": 496.74783120906966,
"children": {
"RLTrainer._checkpoint": {
"total": 0.5762076360006176,
"count": 6,
"self": 0.5762076360006176
}
}
},
"_update_policy": {
"total": 1625.2388456210047,
"count": 1404,
"self": 647.7126751840763,
"children": {
"TorchPPOOptimizer.update": {
"total": 977.5261704369284,
"count": 68343,
"self": 977.5261704369284
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.4979996194597334e-06,
"count": 1,
"self": 1.4979996194597334e-06
},
"TrainerController._save_models": {
"total": 0.09332349099986459,
"count": 1,
"self": 0.001563761999932467,
"children": {
"RLTrainer._checkpoint": {
"total": 0.09175972899993212,
"count": 1,
"self": 0.09175972899993212
}
}
}
}
}
}
}