{ "Huggy": { "checkpoints": [ { "steps": 499990, "file_path": "results/zl-Huggy-01/Huggy/Huggy-499990.onnx", "reward": 4.073455165594052, "creation_time": 1670381824.7145834, "auxillary_file_paths": [ "results/zl-Huggy-01/Huggy/Huggy-499990.pt" ] }, { "steps": 999334, "file_path": "results/zl-Huggy-01/Huggy/Huggy-999334.onnx", "reward": 3.9961099130238673, "creation_time": 1670382439.1763651, "auxillary_file_paths": [ "results/zl-Huggy-01/Huggy/Huggy-999334.pt" ] }, { "steps": 1499923, "file_path": "results/zl-Huggy-01/Huggy/Huggy-1499923.onnx", "reward": 3.973713906847428, "creation_time": 1670383056.6068015, "auxillary_file_paths": [ "results/zl-Huggy-01/Huggy/Huggy-1499923.pt" ] }, { "steps": 1999936, "file_path": "results/zl-Huggy-01/Huggy/Huggy-1999936.onnx", "reward": 4.055368087729629, "creation_time": 1670383688.6106167, "auxillary_file_paths": [ "results/zl-Huggy-01/Huggy/Huggy-1999936.pt" ] }, { "steps": 2000064, "file_path": "results/zl-Huggy-01/Huggy/Huggy-2000064.onnx", "reward": 4.0573191549263745, "creation_time": 1670383688.875483, "auxillary_file_paths": [ "results/zl-Huggy-01/Huggy/Huggy-2000064.pt" ] } ], "final_checkpoint": { "steps": 2000064, "file_path": "results/zl-Huggy-01/Huggy.onnx", "reward": 4.0573191549263745, "creation_time": 1670383688.875483, "auxillary_file_paths": [ "results/zl-Huggy-01/Huggy/Huggy-2000064.pt" ] } }, "metadata": { "stats_format_version": "0.3.0", "mlagents_version": "0.29.0.dev0", "torch_version": "1.8.1+cu102" } }