ppo-Huggy / run_logs /training_status.json
Lelestar's picture
Huggy
4cd9452
raw
history blame
4.38 kB
{
"Huggy": {
"checkpoints": [
{
"steps": 199839,
"file_path": "results/Huggy/Huggy/Huggy-199839.onnx",
"reward": 3.4804463982582092,
"creation_time": 1696425689.8018706,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199839.pt"
]
},
{
"steps": 399870,
"file_path": "results/Huggy/Huggy/Huggy-399870.onnx",
"reward": 3.836338750899784,
"creation_time": 1696425932.4789727,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399870.pt"
]
},
{
"steps": 599944,
"file_path": "results/Huggy/Huggy/Huggy-599944.onnx",
"reward": 3.3618340703748886,
"creation_time": 1696426178.6977715,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599944.pt"
]
},
{
"steps": 799949,
"file_path": "results/Huggy/Huggy/Huggy-799949.onnx",
"reward": 3.712350729459561,
"creation_time": 1696426424.132348,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799949.pt"
]
},
{
"steps": 999845,
"file_path": "results/Huggy/Huggy/Huggy-999845.onnx",
"reward": 4.045841184822288,
"creation_time": 1696426674.1710548,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999845.pt"
]
},
{
"steps": 1199883,
"file_path": "results/Huggy/Huggy/Huggy-1199883.onnx",
"reward": 3.6389196515083313,
"creation_time": 1696426926.3949342,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199883.pt"
]
},
{
"steps": 1399954,
"file_path": "results/Huggy/Huggy/Huggy-1399954.onnx",
"reward": 2.9824755092461905,
"creation_time": 1696427176.433923,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399954.pt"
]
},
{
"steps": 1599916,
"file_path": "results/Huggy/Huggy/Huggy-1599916.onnx",
"reward": 3.80189597747735,
"creation_time": 1696427425.9536889,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599916.pt"
]
},
{
"steps": 1799978,
"file_path": "results/Huggy/Huggy/Huggy-1799978.onnx",
"reward": 3.7803838571519344,
"creation_time": 1696427682.398621,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799978.pt"
]
},
{
"steps": 1999661,
"file_path": "results/Huggy/Huggy/Huggy-1999661.onnx",
"reward": 4.17893797904253,
"creation_time": 1696427939.9351125,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999661.pt"
]
},
{
"steps": 2000411,
"file_path": "results/Huggy/Huggy/Huggy-2000411.onnx",
"reward": 3.9590176235545766,
"creation_time": 1696427940.1750345,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000411.pt"
]
}
],
"final_checkpoint": {
"steps": 2000411,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 3.9590176235545766,
"creation_time": 1696427940.1750345,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000411.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.31.0.dev0",
"torch_version": "1.11.0+cu102"
}
}