ppo-Huggy / run_logs /timers.json
Firemedic15's picture
Huggy
31e1f87 verified
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4017802476882935,
"min": 1.4017802476882935,
"max": 1.427803635597229,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 69097.953125,
"min": 68831.3046875,
"max": 77589.2734375,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 106.06131078224101,
"min": 92.61048689138576,
"max": 425.16101694915255,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 50167.0,
"min": 48990.0,
"max": 50262.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999951.0,
"min": 49537.0,
"max": 1999951.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999951.0,
"min": 49537.0,
"max": 1999951.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.3254199028015137,
"min": 0.1558011919260025,
"max": 2.435394525527954,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1099.923583984375,
"min": 18.22873878479004,
"max": 1270.1156005859375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.534216255768661,
"min": 1.8541804073203323,
"max": 3.896403709968718,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1671.6842889785767,
"min": 216.93910765647888,
"max": 2035.1071266531944,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.534216255768661,
"min": 1.8541804073203323,
"max": 3.896403709968718,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1671.6842889785767,
"min": 216.93910765647888,
"max": 2035.1071266531944,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.017299777227162846,
"min": 0.014470478579157012,
"max": 0.02013711758578817,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.03459955445432569,
"min": 0.028940957158314025,
"max": 0.055187696098194766,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04753213518609603,
"min": 0.020468121115118264,
"max": 0.06057876242945592,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.09506427037219206,
"min": 0.04093624223023653,
"max": 0.16609024927020072,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 4.618298460600005e-06,
"min": 4.618298460600005e-06,
"max": 0.0002952795015734999,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 9.23659692120001e-06,
"min": 9.23659692120001e-06,
"max": 0.00084391336869555,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10153940000000002,
"min": 0.10153940000000002,
"max": 0.1984265,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20307880000000003,
"min": 0.20307880000000003,
"max": 0.5813044500000001,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 8.681606000000012e-05,
"min": 8.681606000000012e-05,
"max": 0.004921482349999999,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00017363212000000023,
"min": 0.00017363212000000023,
"max": 0.014067092055000001,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1717338227",
"python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy2 --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.3.0+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1717340651"
},
"total": 2423.8989084759996,
"count": 1,
"self": 0.44501244299954124,
"children": {
"run_training.setup": {
"total": 0.05947060899995904,
"count": 1,
"self": 0.05947060899995904
},
"TrainerController.start_learning": {
"total": 2423.394425424,
"count": 1,
"self": 4.305843152985744,
"children": {
"TrainerController._reset_env": {
"total": 3.270416828000066,
"count": 1,
"self": 3.270416828000066
},
"TrainerController.advance": {
"total": 2415.7020192390137,
"count": 231831,
"self": 4.5762829829623115,
"children": {
"env_step": {
"total": 1918.5941531880662,
"count": 231831,
"self": 1583.7500014488978,
"children": {
"SubprocessEnvManager._take_step": {
"total": 332.0547609930444,
"count": 231831,
"self": 17.229154540975173,
"children": {
"TorchPolicy.evaluate": {
"total": 314.8256064520692,
"count": 222926,
"self": 314.8256064520692
}
}
},
"workers": {
"total": 2.7893907461240133,
"count": 231831,
"self": 0.0,
"children": {
"worker_root": {
"total": 2416.132290184994,
"count": 231831,
"is_parallel": true,
"self": 1137.2348144871398,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.000929623000047286,
"count": 1,
"is_parallel": true,
"self": 0.00023615600002813153,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006934670000191545,
"count": 2,
"is_parallel": true,
"self": 0.0006934670000191545
}
}
},
"UnityEnvironment.step": {
"total": 0.030351047999943148,
"count": 1,
"is_parallel": true,
"self": 0.00037753499998416373,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00020314699997925345,
"count": 1,
"is_parallel": true,
"self": 0.00020314699997925345
},
"communicator.exchange": {
"total": 0.028918881999970836,
"count": 1,
"is_parallel": true,
"self": 0.028918881999970836
},
"steps_from_proto": {
"total": 0.0008514840000088952,
"count": 1,
"is_parallel": true,
"self": 0.00024648199996590847,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006050020000429868,
"count": 2,
"is_parallel": true,
"self": 0.0006050020000429868
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1278.8974756978541,
"count": 231830,
"is_parallel": true,
"self": 38.82465004792061,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 81.85417095296441,
"count": 231830,
"is_parallel": true,
"self": 81.85417095296441
},
"communicator.exchange": {
"total": 1067.6839733250254,
"count": 231830,
"is_parallel": true,
"self": 1067.6839733250254
},
"steps_from_proto": {
"total": 90.53468137194363,
"count": 231830,
"is_parallel": true,
"self": 32.16958068712802,
"children": {
"_process_rank_one_or_two_observation": {
"total": 58.365100684815616,
"count": 463660,
"is_parallel": true,
"self": 58.365100684815616
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 492.53158306798525,
"count": 231831,
"self": 6.312307058952001,
"children": {
"process_trajectory": {
"total": 151.21788829103366,
"count": 231831,
"self": 149.9625106490338,
"children": {
"RLTrainer._checkpoint": {
"total": 1.2553776419998712,
"count": 10,
"self": 1.2553776419998712
}
}
},
"_update_policy": {
"total": 335.0013877179996,
"count": 96,
"self": 271.60439146099657,
"children": {
"TorchPPOOptimizer.update": {
"total": 63.396996257003025,
"count": 2880,
"self": 63.396996257003025
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.4720003491675016e-06,
"count": 1,
"self": 1.4720003491675016e-06
},
"TrainerController._save_models": {
"total": 0.11614473200006614,
"count": 1,
"self": 0.002182636999805254,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11396209500026089,
"count": 1,
"self": 0.11396209500026089
}
}
}
}
}
}
}