ppo-Huggy / run_logs /training_status.json
joshbz's picture
Huggy
d782d95 verified
raw
history blame contribute delete
No virus
4.4 kB
{
"Huggy": {
"checkpoints": [
{
"steps": 199934,
"file_path": "results/Huggy2/Huggy/Huggy-199934.onnx",
"reward": 3.4942999941403747,
"creation_time": 1719952979.2986002,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-199934.pt"
]
},
{
"steps": 399826,
"file_path": "results/Huggy2/Huggy/Huggy-399826.onnx",
"reward": 3.5624916513760883,
"creation_time": 1719953208.8541129,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-399826.pt"
]
},
{
"steps": 599981,
"file_path": "results/Huggy2/Huggy/Huggy-599981.onnx",
"reward": 3.4186143227245496,
"creation_time": 1719953437.8381681,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-599981.pt"
]
},
{
"steps": 799931,
"file_path": "results/Huggy2/Huggy/Huggy-799931.onnx",
"reward": 3.8894465370820117,
"creation_time": 1719953666.0088046,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-799931.pt"
]
},
{
"steps": 999983,
"file_path": "results/Huggy2/Huggy/Huggy-999983.onnx",
"reward": 3.921925241811366,
"creation_time": 1719953901.2991803,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-999983.pt"
]
},
{
"steps": 1199967,
"file_path": "results/Huggy2/Huggy/Huggy-1199967.onnx",
"reward": 3.6672738355559273,
"creation_time": 1719954135.2872849,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1199967.pt"
]
},
{
"steps": 1399934,
"file_path": "results/Huggy2/Huggy/Huggy-1399934.onnx",
"reward": 3.921330222716698,
"creation_time": 1719954369.7131164,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1399934.pt"
]
},
{
"steps": 1599970,
"file_path": "results/Huggy2/Huggy/Huggy-1599970.onnx",
"reward": 3.8724617221469964,
"creation_time": 1719954602.6336875,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1599970.pt"
]
},
{
"steps": 1799969,
"file_path": "results/Huggy2/Huggy/Huggy-1799969.onnx",
"reward": 4.026444941970474,
"creation_time": 1719954837.9440691,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1799969.pt"
]
},
{
"steps": 1999958,
"file_path": "results/Huggy2/Huggy/Huggy-1999958.onnx",
"reward": 3.721830139725895,
"creation_time": 1719955072.8875318,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1999958.pt"
]
},
{
"steps": 2000042,
"file_path": "results/Huggy2/Huggy/Huggy-2000042.onnx",
"reward": 3.7467932363351184,
"creation_time": 1719955073.0188763,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-2000042.pt"
]
}
],
"final_checkpoint": {
"steps": 2000042,
"file_path": "results/Huggy2/Huggy.onnx",
"reward": 3.7467932363351184,
"creation_time": 1719955073.0188763,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-2000042.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "1.1.0.dev0",
"torch_version": "2.3.0+cu121"
}
}