philippds's picture
Upload 13 files
8a58754 verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 1.4533461332321167,
"min": 1.4189385175704956,
"max": 1.4572324752807617,
"count": 200
},
"Agent.Policy.Entropy.sum": {
"value": 9068.8798828125,
"min": 7246.6962890625,
"max": 9839.8515625,
"count": 200
},
"Agent.DroneBasedReforestation.TreeDropCount.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.TreeDropCount.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.RechargeEnergyCount.mean": {
"value": 129.76190476190476,
"min": 0.0,
"max": 417.93333333333334,
"count": 200
},
"Agent.DroneBasedReforestation.RechargeEnergyCount.sum": {
"value": 2725.0,
"min": 0.0,
"max": 6269.0,
"count": 200
},
"Agent.DroneBasedReforestation.SaveLocationCount.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.SaveLocationCount.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.OutofEnergyCount.mean": {
"value": 0.6190476190476191,
"min": 0.4444444444444444,
"max": 0.8,
"count": 200
},
"Agent.DroneBasedReforestation.OutofEnergyCount.sum": {
"value": 13.0,
"min": 7.0,
"max": 15.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.EpisodeLength.mean": {
"value": 346.3333333333333,
"min": 276.14285714285717,
"max": 399.0,
"count": 200
},
"Agent.Environment.EpisodeLength.sum": {
"value": 6234.0,
"min": 4956.0,
"max": 6765.0,
"count": 200
},
"Agent.Step.mean": {
"value": 1199610.0,
"min": 5600.0,
"max": 1199610.0,
"count": 200
},
"Agent.Step.sum": {
"value": 1199610.0,
"min": 5600.0,
"max": 1199610.0,
"count": 200
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.18385379016399384,
"min": 0.0292599406093359,
"max": 1.0435168743133545,
"count": 200
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 3.125514507293701,
"min": 0.43889909982681274,
"max": 15.652752876281738,
"count": 200
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.8692796230316162,
"min": -0.172968327999115,
"max": 3.2063863277435303,
"count": 200
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 14.777753829956055,
"min": -2.940021514892578,
"max": 67.28934478759766,
"count": 200
},
"Agent.Environment.CumulativeReward.mean": {
"value": 10.825707225238576,
"min": -1.132599985599518,
"max": 30.65860773701417,
"count": 200
},
"Agent.Environment.CumulativeReward.sum": {
"value": 184.0370228290558,
"min": -16.988999783992767,
"max": 582.5135470032692,
"count": 200
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.6476288959383965,
"min": 0.0,
"max": 14.209401581022474,
"count": 200
},
"Agent.Policy.CuriosityReward.sum": {
"value": 11.00969123095274,
"min": 0.0,
"max": 255.76922845840454,
"count": 200
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 9.74313505432185,
"min": -1.0193402210871378,
"max": 27.592743119126872,
"count": 200
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 165.63329592347145,
"min": -15.290103316307068,
"max": 524.2621192634106,
"count": 200
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 200
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 200
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.02661372983129695,
"min": 0.014345233794301748,
"max": 0.03552510698015491,
"count": 139
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.02661372983129695,
"min": 0.014345233794301748,
"max": 0.03552510698015491,
"count": 139
},
"Agent.Losses.ValueLoss.mean": {
"value": 7.1942349672317505,
"min": 0.0016999954968923703,
"max": 21.090327545448584,
"count": 139
},
"Agent.Losses.ValueLoss.sum": {
"value": 7.1942349672317505,
"min": 0.0016999954968923703,
"max": 21.090327545448584,
"count": 139
},
"Agent.Policy.LearningRate.mean": {
"value": 1.660599446499987e-06,
"min": 1.660599446499987e-06,
"max": 0.0002979000007,
"count": 139
},
"Agent.Policy.LearningRate.sum": {
"value": 1.660599446499987e-06,
"min": 1.660599446499987e-06,
"max": 0.0002979000007,
"count": 139
},
"Agent.Policy.Epsilon.mean": {
"value": 0.1005535,
"min": 0.1005535,
"max": 0.1993,
"count": 139
},
"Agent.Policy.Epsilon.sum": {
"value": 0.1005535,
"min": 0.1005535,
"max": 0.1993,
"count": 139
},
"Agent.Policy.Beta.mean": {
"value": 3.761964999999979e-05,
"min": 3.761964999999979e-05,
"max": 0.00496507,
"count": 139
},
"Agent.Policy.Beta.sum": {
"value": 3.761964999999979e-05,
"min": 3.761964999999979e-05,
"max": 0.00496507,
"count": 139
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.017956590512767434,
"min": 0.017066249158233404,
"max": 0.5835290277997652,
"count": 139
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.017956590512767434,
"min": 0.017066249158233404,
"max": 0.5835290277997652,
"count": 139
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 2.0348302821318307,
"min": 1.9834658404191334,
"max": 3.310828596353531,
"count": 139
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 2.0348302821318307,
"min": 1.9834658404191334,
"max": 3.310828596353531,
"count": 139
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1717574791",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_3_task_2_run_id_2_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_3_task_2_run_id_2_train --base-port 5007",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1717579534"
},
"total": 4742.8268369,
"count": 1,
"self": 0.32466740000018035,
"children": {
"run_training.setup": {
"total": 0.07259170000000004,
"count": 1,
"self": 0.07259170000000004
},
"TrainerController.start_learning": {
"total": 4742.4295778,
"count": 1,
"self": 8.336508300056266,
"children": {
"TrainerController._reset_env": {
"total": 2.2312725,
"count": 1,
"self": 2.2312725
},
"TrainerController.advance": {
"total": 4731.6206696999425,
"count": 401158,
"self": 7.422396399857462,
"children": {
"env_step": {
"total": 4724.198273300085,
"count": 401158,
"self": 1947.7122418999506,
"children": {
"SubprocessEnvManager._take_step": {
"total": 2771.474913300064,
"count": 401158,
"self": 13.969147400131988,
"children": {
"TorchPolicy.evaluate": {
"total": 2757.505765899932,
"count": 400270,
"self": 2757.505765899932
}
}
},
"workers": {
"total": 5.011118100070284,
"count": 401158,
"self": 0.0,
"children": {
"worker_root": {
"total": 4732.350145099952,
"count": 401158,
"is_parallel": true,
"self": 3067.838198099922,
"children": {
"steps_from_proto": {
"total": 0.006916900000000004,
"count": 1,
"is_parallel": true,
"self": 0.00010650000000000936,
"children": {
"_process_maybe_compressed_observation": {
"total": 0.006762600000000063,
"count": 2,
"is_parallel": true,
"self": 3.790000000014615e-05,
"children": {
"_observation_to_np_array": {
"total": 0.006724699999999917,
"count": 3,
"is_parallel": true,
"self": 3.399999999986747e-05,
"children": {
"process_pixels": {
"total": 0.006690700000000049,
"count": 3,
"is_parallel": true,
"self": 0.000247200000000003,
"children": {
"image_decompress": {
"total": 0.006443500000000046,
"count": 3,
"is_parallel": true,
"self": 0.006443500000000046
}
}
}
}
}
}
},
"_process_rank_one_or_two_observation": {
"total": 4.779999999993123e-05,
"count": 2,
"is_parallel": true,
"self": 4.779999999993123e-05
}
}
},
"UnityEnvironment.step": {
"total": 1664.5050301000297,
"count": 401158,
"is_parallel": true,
"self": 26.92025999966586,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 26.473373000047772,
"count": 401158,
"is_parallel": true,
"self": 26.473373000047772
},
"communicator.exchange": {
"total": 1405.283343300271,
"count": 401158,
"is_parallel": true,
"self": 1405.283343300271
},
"steps_from_proto": {
"total": 205.8280538000449,
"count": 401158,
"is_parallel": true,
"self": 41.25461419988886,
"children": {
"_process_maybe_compressed_observation": {
"total": 146.472787500226,
"count": 802316,
"is_parallel": true,
"self": 11.28791419999743,
"children": {
"_observation_to_np_array": {
"total": 135.18487330022856,
"count": 1204062,
"is_parallel": true,
"self": 11.567830200627753,
"children": {
"process_pixels": {
"total": 123.6170430996008,
"count": 1204062,
"is_parallel": true,
"self": 58.48534459972299,
"children": {
"image_decompress": {
"total": 65.13169849987781,
"count": 1204062,
"is_parallel": true,
"self": 65.13169849987781
}
}
}
}
}
}
},
"_process_rank_one_or_two_observation": {
"total": 18.100652099930034,
"count": 802316,
"is_parallel": true,
"self": 18.100652099930034
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 3.49000001733657e-05,
"count": 1,
"self": 3.49000001733657e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 4736.830238500035,
"count": 228938,
"is_parallel": true,
"self": 9.691872500030513,
"children": {
"process_trajectory": {
"total": 3745.0100708000027,
"count": 228938,
"is_parallel": true,
"self": 3744.3179560000026,
"children": {
"RLTrainer._checkpoint": {
"total": 0.6921147999998993,
"count": 2,
"is_parallel": true,
"self": 0.6921147999998993
}
}
},
"_update_policy": {
"total": 982.128295200002,
"count": 139,
"is_parallel": true,
"self": 654.2106892000122,
"children": {
"TorchPPOOptimizer.update": {
"total": 327.9176059999897,
"count": 3357,
"is_parallel": true,
"self": 327.9176059999897
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.24109240000052523,
"count": 1,
"self": 0.006454300000768853,
"children": {
"RLTrainer._checkpoint": {
"total": 0.23463809999975638,
"count": 1,
"self": 0.23463809999975638
}
}
}
}
}
}
}