ppo-Huggy / run_logs /training_status.json
MarioLomby's picture
Huggy
afd301b
raw
history blame contribute delete
No virus
4.38 kB
{
"Huggy": {
"checkpoints": [
{
"steps": 199902,
"file_path": "results/Huggy/Huggy/Huggy-199902.onnx",
"reward": 3.2982311826199293,
"creation_time": 1675454556.8425992,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199902.pt"
]
},
{
"steps": 399941,
"file_path": "results/Huggy/Huggy/Huggy-399941.onnx",
"reward": 3.448641276359558,
"creation_time": 1675454823.4356344,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399941.pt"
]
},
{
"steps": 599956,
"file_path": "results/Huggy/Huggy/Huggy-599956.onnx",
"reward": 3.5397546042998633,
"creation_time": 1675455090.4352844,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599956.pt"
]
},
{
"steps": 799979,
"file_path": "results/Huggy/Huggy/Huggy-799979.onnx",
"reward": 3.872860514832114,
"creation_time": 1675455357.2573364,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799979.pt"
]
},
{
"steps": 999943,
"file_path": "results/Huggy/Huggy/Huggy-999943.onnx",
"reward": 3.9588643918037416,
"creation_time": 1675455632.1974297,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999943.pt"
]
},
{
"steps": 1199934,
"file_path": "results/Huggy/Huggy/Huggy-1199934.onnx",
"reward": 4.0540529023855925,
"creation_time": 1675455903.5731132,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199934.pt"
]
},
{
"steps": 1399955,
"file_path": "results/Huggy/Huggy/Huggy-1399955.onnx",
"reward": 3.8562845582824674,
"creation_time": 1675456168.4676876,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399955.pt"
]
},
{
"steps": 1599983,
"file_path": "results/Huggy/Huggy/Huggy-1599983.onnx",
"reward": 3.7040274952359766,
"creation_time": 1675456443.927257,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599983.pt"
]
},
{
"steps": 1799940,
"file_path": "results/Huggy/Huggy/Huggy-1799940.onnx",
"reward": 3.6232298458064043,
"creation_time": 1675456722.2061744,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799940.pt"
]
},
{
"steps": 1999973,
"file_path": "results/Huggy/Huggy/Huggy-1999973.onnx",
"reward": 4.138221814277324,
"creation_time": 1675457005.6243575,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999973.pt"
]
},
{
"steps": 2000023,
"file_path": "results/Huggy/Huggy/Huggy-2000023.onnx",
"reward": 4.108608963588874,
"creation_time": 1675457005.8186169,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000023.pt"
]
}
],
"final_checkpoint": {
"steps": 2000023,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 4.108608963588874,
"creation_time": 1675457005.8186169,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000023.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.29.0.dev0",
"torch_version": "1.8.1+cu102"
}
}