{ "name": "root", "gauges": { "Agent.Policy.Entropy.mean": { "value": 1.3833463191986084, "min": 1.3833130598068237, "max": 1.4281889200210571, "count": 200 }, "Agent.Policy.Entropy.sum": { "value": 8341.578125, "min": 6968.845703125, "max": 9999.2880859375, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.mean": { "value": 0.967741935483871, "min": 0.4, "max": 1.0, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.sum": { "value": 90.0, "min": 6.0, "max": 153.0, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.mean": { "value": 11.35483870967742, "min": 5.037037037037037, "max": 44.8, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.sum": { "value": 1056.0, "min": 100.0, "max": 1767.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.mean": { "value": 0.021505376344086023, "min": 0.0, "max": 0.4666666666666667, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.sum": { "value": 2.0, "min": 0.0, "max": 11.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": { "value": 52.0167432600452, "min": 13.662641604741415, "max": 65.12439632415771, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": { "value": 4837.557123184204, "min": 327.90339851379395, "max": 7769.0461683273315, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": { "value": 3.928516927585807, "min": 0.49648803658783436, "max": 7.833532768182265, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": { "value": 365.35207426548004, "min": 11.915712878108025, "max": 584.8969851732254, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": { "value": 1.262566012720908, "min": 0.22728414088487625, "max": 2.116979764567481, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": { "value": 117.41863918304443, "min": 5.45481938123703, "max": 188.2134671807289, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": { "value": 0.12625660202516023, "min": 0.022728413964311283, "max": 0.21169797579447427, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": { "value": 11.741863988339901, "min": 0.5454819351434708, "max": 18.821346774697304, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": { "value": 70.43514513008056, "min": 24.010529394502992, "max": 76.13874863783518, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": { "value": 6550.468497097492, "min": 524.863603591919, "max": 10701.750457644463, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.EpisodeLength.mean": { "value": 63.83870967741935, "min": 36.73913043478261, "max": 359.4, "count": 200 }, "Agent.Environment.EpisodeLength.sum": { "value": 5937.0, "min": 4863.0, "max": 7065.0, "count": 200 }, "Agent.Step.mean": { "value": 1199712.0, "min": 5673.0, "max": 1199712.0, "count": 200 }, "Agent.Step.sum": { "value": 1199712.0, "min": 5673.0, "max": 1199712.0, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.mean": { "value": 0.2415272742509842, "min": 0.01319605577737093, "max": 1.1567590236663818, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.sum": { "value": 22.4620361328125, "min": 0.39588168263435364, "max": 35.6578369140625, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.mean": { "value": 53.40044403076172, "min": 0.05967150628566742, "max": 56.91693878173828, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.sum": { "value": 4966.2412109375, "min": 1.4917876720428467, "max": 7830.58544921875, "count": 200 }, "Agent.Environment.CumulativeReward.mean": { "value": 101.52704759863435, "min": 38.354348639026284, "max": 106.32891958289676, "count": 200 }, "Agent.Environment.CumulativeReward.sum": { "value": 9442.015426672995, "min": 613.6695782244205, "max": 16037.046478271484, "count": 200 }, "Agent.Policy.CuriosityReward.mean": { "value": 0.2271482430158123, "min": 0.0, "max": 5.984522547040667, "count": 200 }, "Agent.Policy.CuriosityReward.sum": { "value": 21.124786600470543, "min": 0.0, "max": 209.45828914642334, "count": 200 }, "Agent.Policy.ExtrinsicReward.mean": { "value": 91.37434429166618, "min": 34.518917528912425, "max": 95.69603141148885, "count": 200 }, "Agent.Policy.ExtrinsicReward.sum": { "value": 8497.814019124955, "min": 552.3026804625988, "max": 14433.342140197754, "count": 200 }, "Agent.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.Losses.PolicyLoss.mean": { "value": 0.01996750299197932, "min": 0.014557186746969819, "max": 0.02999215660383925, "count": 138 }, "Agent.Losses.PolicyLoss.sum": { "value": 0.01996750299197932, "min": 0.014557186746969819, "max": 0.02999215660383925, "count": 138 }, "Agent.Losses.ValueLoss.mean": { "value": 98.03205680847168, "min": 31.185107072194416, "max": 143.23234367370605, "count": 138 }, "Agent.Losses.ValueLoss.sum": { "value": 98.03205680847168, "min": 31.185107072194416, "max": 143.23234367370605, "count": 138 }, "Agent.Policy.LearningRate.mean": { "value": 1.7303494232500109e-06, "min": 1.7303494232500109e-06, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.LearningRate.sum": { "value": 1.7303494232500109e-06, "min": 1.7303494232500109e-06, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.Epsilon.mean": { "value": 0.10057675, "min": 0.10057675, "max": 0.1992545, "count": 138 }, "Agent.Policy.Epsilon.sum": { "value": 0.10057675, "min": 0.10057675, "max": 0.1992545, "count": 138 }, "Agent.Policy.Beta.mean": { "value": 3.877982500000019e-05, "min": 3.877982500000019e-05, "max": 0.00496279955, "count": 138 }, "Agent.Policy.Beta.sum": { "value": 3.877982500000019e-05, "min": 3.877982500000019e-05, "max": 0.00496279955, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.mean": { "value": 0.029792719598238666, "min": 0.01700370052518944, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.sum": { "value": 0.029792719598238666, "min": 0.01700370052518944, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.mean": { "value": 2.222791443268458, "min": 2.222791443268458, "max": 3.353024572134018, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.sum": { "value": 2.222791443268458, "min": 2.222791443268458, "max": 3.353024572134018, "count": 138 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1717395331", "python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_9_task_3_run_id_1_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_9_task_3_run_id_1_train --base-port 5007", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.7.1+cu110", "numpy_version": "1.21.0", "end_time_seconds": "1717399460" }, "total": 4128.9967185, "count": 1, "self": 0.7428867999997237, "children": { "run_training.setup": { "total": 0.054176800000000025, "count": 1, "self": 0.054176800000000025 }, "TrainerController.start_learning": { "total": 4128.1996549, "count": 1, "self": 7.166977399929237, "children": { "TrainerController._reset_env": { "total": 2.0551066999999996, "count": 1, "self": 2.0551066999999996 }, "TrainerController.advance": { "total": 4118.77019670007, "count": 403908, "self": 6.831238199963082, "children": { "env_step": { "total": 4111.938958500107, "count": 403908, "self": 2192.994379500303, "children": { "SubprocessEnvManager._take_step": { "total": 1914.6036500998744, "count": 403908, "self": 12.136192299768936, "children": { "TorchPolicy.evaluate": { "total": 1902.4674578001054, "count": 400304, "self": 1902.4674578001054 } } }, "workers": { "total": 4.340928899929992, "count": 403908, "self": 0.0, "children": { "worker_root": { "total": 4119.652724599985, "count": 403908, "is_parallel": true, "self": 2157.5229497999476, "children": { "steps_from_proto": { "total": 0.006367099999999848, "count": 1, "is_parallel": true, "self": 0.0001040999999997183, "children": { "_process_maybe_compressed_observation": { "total": 0.006217600000000045, "count": 2, "is_parallel": true, "self": 3.260000000016028e-05, "children": { "_observation_to_np_array": { "total": 0.006184999999999885, "count": 3, "is_parallel": true, "self": 3.100000000011427e-05, "children": { "process_pixels": { "total": 0.006153999999999771, "count": 3, "is_parallel": true, "self": 0.00023029999999990558, "children": { "image_decompress": { "total": 0.005923699999999865, "count": 3, "is_parallel": true, "self": 0.005923699999999865 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 4.540000000008426e-05, "count": 2, "is_parallel": true, "self": 4.540000000008426e-05 } } }, "UnityEnvironment.step": { "total": 1962.1234077000374, "count": 403908, "is_parallel": true, "self": 17.9943053001914, "children": { "UnityEnvironment._generate_step_input": { "total": 17.937010500057944, "count": 403908, "is_parallel": true, "self": 17.937010500057944 }, "communicator.exchange": { "total": 1772.1705877998884, "count": 403908, "is_parallel": true, "self": 1772.1705877998884 }, "steps_from_proto": { "total": 154.02150409989957, "count": 403908, "is_parallel": true, "self": 30.778080100048072, "children": { "_process_maybe_compressed_observation": { "total": 110.84091120002984, "count": 807816, "is_parallel": true, "self": 8.402484899910093, "children": { "_observation_to_np_array": { "total": 102.43842630011974, "count": 1212324, "is_parallel": true, "self": 8.140582299767715, "children": { "process_pixels": { "total": 94.29784400035203, "count": 1212324, "is_parallel": true, "self": 43.510258600415824, "children": { "image_decompress": { "total": 50.787585399936205, "count": 1212324, "is_parallel": true, "self": 50.787585399936205 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 12.402512799821672, "count": 807816, "is_parallel": true, "self": 12.402512799821672 } } } } } } } } } } } } }, "trainer_threads": { "total": 3.310000010969816e-05, "count": 1, "self": 3.310000010969816e-05, "children": { "thread_root": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "trainer_advance": { "total": 4123.99575400007, "count": 206462, "is_parallel": true, "self": 6.190365099997507, "children": { "process_trajectory": { "total": 3414.869868700075, "count": 206462, "is_parallel": true, "self": 3414.4624559000754, "children": { "RLTrainer._checkpoint": { "total": 0.4074127999997472, "count": 2, "is_parallel": true, "self": 0.4074127999997472 } } }, "_update_policy": { "total": 702.9355201999971, "count": 138, "is_parallel": true, "self": 475.6080876000008, "children": { "TorchPPOOptimizer.update": { "total": 227.3274325999963, "count": 3357, "is_parallel": true, "self": 227.3274325999963 } } } } } } } } }, "TrainerController._save_models": { "total": 0.20734100000026956, "count": 1, "self": 0.007222400000500784, "children": { "RLTrainer._checkpoint": { "total": 0.20011859999976878, "count": 1, "self": 0.20011859999976878 } } } } } } }