philippds's picture
Upload 9 files
e086ef0 verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.9181978702545166,
"min": 3.8353826999664307,
"max": 4.385557174682617,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 21651.9609375,
"min": 7004.57666015625,
"max": 27602.22265625,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8666667496716535,
"min": 0.5777778308838606,
"max": 0.9000000953674316,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 23.400002241134644,
"min": 7.700000762939453,
"max": 32.20000338554382,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 2.152441685018992,
"min": 0.8010983359482553,
"max": 15.489594587573299,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 58.11592549551278,
"min": 11.439244419336319,
"max": 418.21905386447906,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 304.5739384757148,
"min": 26.05030299226443,
"max": 468.1476593017578,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 8223.4963388443,
"min": 937.8109077215195,
"max": 8394.559112548828,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 64.0783942187274,
"min": 20.735190947850544,
"max": 117.72799767388238,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 1730.1166439056396,
"min": 435.9099578857422,
"max": 1998.507007598877,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 35.4022040632036,
"min": 11.265056636598375,
"max": 57.77317428588867,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 955.8595097064972,
"min": 220.9379529953003,
"max": 1043.3600313663483,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 204.66666666666666,
"min": 93.75,
"max": 497.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 5526.0,
"min": 1719.0,
"max": 6831.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449812.0,
"min": 4398.0,
"max": 449812.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449812.0,
"min": 4398.0,
"max": 449812.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.0032545796129852533,
"min": -0.04245872423052788,
"max": 0.11642169952392578,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 0.06834617257118225,
"min": -0.9340919256210327,
"max": 3.118682861328125,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 101.64611053466797,
"min": 3.4324796199798584,
"max": 117.11848449707031,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 2134.568359375,
"min": 85.8119888305664,
"max": 3188.197509765625,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 388.0023089817592,
"min": 76.50909111022949,
"max": 603.0836317274305,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 8148.048488616943,
"min": 1912.7272777557373,
"max": 8148.048488616943,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.027668442399728866,
"min": 0.02111277556805699,
"max": 0.18603225509551438,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.5810372903943062,
"min": 0.5008933562785387,
"max": 4.0927096121013165,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 388.0023089817592,
"min": 76.50909111022949,
"max": 603.0836317274305,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 8148.048488616943,
"min": 1912.7272777557373,
"max": 8148.048488616943,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.07081948476148817,
"min": 0.051579486404856045,
"max": 0.08544279739522097,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.14163896952297633,
"min": 0.051579486404856045,
"max": 0.15435190841607335,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 919.6023082684374,
"min": 36.81519386224579,
"max": 919.6023082684374,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 1839.2046165368747,
"min": 36.81519386224579,
"max": 1839.2046165368747,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.7790994069999974e-06,
"min": 1.7790994069999974e-06,
"max": 0.00029834400055199994,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 3.5581988139999947e-06,
"min": 3.5581988139999947e-06,
"max": 0.0005912520029159999,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.100593,
"min": 0.100593,
"max": 0.19944800000000004,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.201186,
"min": 0.10142999999999998,
"max": 0.3970840000000001,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 6.924069999999992e-05,
"min": 6.924069999999992e-05,
"max": 0.009944855199999998,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00013848139999999984,
"min": 0.00013848139999999984,
"max": 0.019708691599999995,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.005482690535463514,
"min": 0.004134121945955687,
"max": 0.3886077385722545,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.010965381070927028,
"min": 0.004134121945955687,
"max": 0.3886077385722545,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.85200771986676,
"min": 3.732027288467165,
"max": 4.39014532691554,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.70401543973352,
"min": 3.732027288467165,
"max": 8.739337876731273,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716727898",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_8_task_1_run_id_1_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_8_task_1_run_id_1_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716729806"
},
"total": 1907.8616203000001,
"count": 1,
"self": 0.2091005000002042,
"children": {
"run_training.setup": {
"total": 0.05204299999999995,
"count": 1,
"self": 0.05204299999999995
},
"TrainerController.start_learning": {
"total": 1907.6004768,
"count": 1,
"self": 1.3724266000042462,
"children": {
"TrainerController._reset_env": {
"total": 2.2182755,
"count": 1,
"self": 2.2182755
},
"TrainerController.advance": {
"total": 1903.8876972999958,
"count": 50157,
"self": 1.1044949999766231,
"children": {
"env_step": {
"total": 1902.7832023000192,
"count": 50157,
"self": 1589.0755829000168,
"children": {
"SubprocessEnvManager._take_step": {
"total": 313.19396319999316,
"count": 50157,
"self": 2.029523899969149,
"children": {
"TorchPolicy.evaluate": {
"total": 311.164439300024,
"count": 50157,
"self": 311.164439300024
}
}
},
"workers": {
"total": 0.5136562000090978,
"count": 50157,
"self": 0.0,
"children": {
"worker_root": {
"total": 1904.7780526000035,
"count": 50157,
"is_parallel": true,
"self": 385.78522210001825,
"children": {
"steps_from_proto": {
"total": 0.0002884999999999138,
"count": 1,
"is_parallel": true,
"self": 0.00012699999999998823,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0001614999999999256,
"count": 2,
"is_parallel": true,
"self": 0.0001614999999999256
}
}
},
"UnityEnvironment.step": {
"total": 1518.9925419999854,
"count": 50157,
"is_parallel": true,
"self": 4.3082015999791565,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 5.646964899999924,
"count": 50157,
"is_parallel": true,
"self": 5.646964899999924
},
"communicator.exchange": {
"total": 1497.4812175999834,
"count": 50157,
"is_parallel": true,
"self": 1497.4812175999834
},
"steps_from_proto": {
"total": 11.556157900023079,
"count": 50157,
"is_parallel": true,
"self": 6.010886400029484,
"children": {
"_process_rank_one_or_two_observation": {
"total": 5.545271499993595,
"count": 100314,
"is_parallel": true,
"self": 5.545271499993595
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.7199999976801337e-05,
"count": 1,
"self": 2.7199999976801337e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1904.0746734999818,
"count": 92880,
"is_parallel": true,
"self": 3.372148499981222,
"children": {
"process_trajectory": {
"total": 1474.4498105000011,
"count": 92880,
"is_parallel": true,
"self": 1474.4498105000011
},
"_update_policy": {
"total": 426.25271449999946,
"count": 148,
"is_parallel": true,
"self": 209.94881250000054,
"children": {
"TorchPPOOptimizer.update": {
"total": 216.3039019999989,
"count": 10284,
"is_parallel": true,
"self": 216.3039019999989
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.12205020000010336,
"count": 1,
"self": 0.005611200000203098,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11643899999990026,
"count": 1,
"self": 0.11643899999990026
}
}
}
}
}
}
}