ppo-Huggy / run_logs /training_status.json
heziyevv's picture
Huggy
098e0fe
raw
history blame
4.38 kB
{
"Huggy": {
"checkpoints": [
{
"steps": 199995,
"file_path": "results/Huggy/Huggy/Huggy-199995.onnx",
"reward": 3.3686767528134007,
"creation_time": 1678950131.5015268,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199995.pt"
]
},
{
"steps": 399885,
"file_path": "results/Huggy/Huggy/Huggy-399885.onnx",
"reward": 3.6761927760564364,
"creation_time": 1678950361.3844588,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399885.pt"
]
},
{
"steps": 599963,
"file_path": "results/Huggy/Huggy/Huggy-599963.onnx",
"reward": 3.3405493895212808,
"creation_time": 1678950593.7370954,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599963.pt"
]
},
{
"steps": 799986,
"file_path": "results/Huggy/Huggy/Huggy-799986.onnx",
"reward": 3.6978548798464277,
"creation_time": 1678950826.032446,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799986.pt"
]
},
{
"steps": 999888,
"file_path": "results/Huggy/Huggy/Huggy-999888.onnx",
"reward": 3.751657053937007,
"creation_time": 1678951060.233893,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999888.pt"
]
},
{
"steps": 1199985,
"file_path": "results/Huggy/Huggy/Huggy-1199985.onnx",
"reward": 3.9383012832157194,
"creation_time": 1678951293.275635,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199985.pt"
]
},
{
"steps": 1399952,
"file_path": "results/Huggy/Huggy/Huggy-1399952.onnx",
"reward": 3.6296459990265095,
"creation_time": 1678951522.6272588,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399952.pt"
]
},
{
"steps": 1599878,
"file_path": "results/Huggy/Huggy/Huggy-1599878.onnx",
"reward": 3.4336966745469764,
"creation_time": 1678951757.0576546,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599878.pt"
]
},
{
"steps": 1799268,
"file_path": "results/Huggy/Huggy/Huggy-1799268.onnx",
"reward": 3.6458883172935908,
"creation_time": 1678951988.8274212,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799268.pt"
]
},
{
"steps": 1999341,
"file_path": "results/Huggy/Huggy/Huggy-1999341.onnx",
"reward": 3.5990183220969305,
"creation_time": 1678952220.7693527,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999341.pt"
]
},
{
"steps": 2000091,
"file_path": "results/Huggy/Huggy/Huggy-2000091.onnx",
"reward": 3.204075311359606,
"creation_time": 1678952220.9167745,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000091.pt"
]
}
],
"final_checkpoint": {
"steps": 2000091,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 3.204075311359606,
"creation_time": 1678952220.9167745,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000091.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.31.0.dev0",
"torch_version": "1.11.0+cu102"
}
}