ppo-Huggy / run_logs /training_status.json
Bryan5's picture
Huggy
d160b82
raw
history blame contribute delete
No virus
4.38 kB
{
"Huggy": {
"checkpoints": [
{
"steps": 199980,
"file_path": "results/Huggy/Huggy/Huggy-199980.onnx",
"reward": 2.836747789177401,
"creation_time": 1691482916.6079817,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199980.pt"
]
},
{
"steps": 399929,
"file_path": "results/Huggy/Huggy/Huggy-399929.onnx",
"reward": 3.4224201942745007,
"creation_time": 1691483179.354004,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399929.pt"
]
},
{
"steps": 599978,
"file_path": "results/Huggy/Huggy/Huggy-599978.onnx",
"reward": 3.664274439116977,
"creation_time": 1691483442.8623264,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599978.pt"
]
},
{
"steps": 799942,
"file_path": "results/Huggy/Huggy/Huggy-799942.onnx",
"reward": 3.512777098724919,
"creation_time": 1691483714.8426785,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799942.pt"
]
},
{
"steps": 999984,
"file_path": "results/Huggy/Huggy/Huggy-999984.onnx",
"reward": 3.5100431387451874,
"creation_time": 1691483985.1270337,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999984.pt"
]
},
{
"steps": 1199915,
"file_path": "results/Huggy/Huggy/Huggy-1199915.onnx",
"reward": 3.2000963389873505,
"creation_time": 1691484256.1240864,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199915.pt"
]
},
{
"steps": 1399956,
"file_path": "results/Huggy/Huggy/Huggy-1399956.onnx",
"reward": 3.424461195724351,
"creation_time": 1691484520.6845741,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399956.pt"
]
},
{
"steps": 1599963,
"file_path": "results/Huggy/Huggy/Huggy-1599963.onnx",
"reward": 3.6753125275884355,
"creation_time": 1691484789.5009358,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599963.pt"
]
},
{
"steps": 1799944,
"file_path": "results/Huggy/Huggy/Huggy-1799944.onnx",
"reward": 3.5871960087256,
"creation_time": 1691485056.286318,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799944.pt"
]
},
{
"steps": 1999828,
"file_path": "results/Huggy/Huggy/Huggy-1999828.onnx",
"reward": 3.3589174844694476,
"creation_time": 1691485317.6302252,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999828.pt"
]
},
{
"steps": 2000578,
"file_path": "results/Huggy/Huggy/Huggy-2000578.onnx",
"reward": 3.314090322021028,
"creation_time": 1691485317.8006396,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000578.pt"
]
}
],
"final_checkpoint": {
"steps": 2000578,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 3.314090322021028,
"creation_time": 1691485317.8006396,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000578.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.31.0.dev0",
"torch_version": "1.11.0+cu102"
}
}