File size: 152 Bytes
d7be120 |
1 |
{"mean_reward": 599.5, "std_reward": 239.7754157539926, "is_deterministic": false, "n_eval_episodes": 10, "eval_datetime": "2023-09-14T17:01:36.346105"} |
d7be120 |
1 |
{"mean_reward": 599.5, "std_reward": 239.7754157539926, "is_deterministic": false, "n_eval_episodes": 10, "eval_datetime": "2023-09-14T17:01:36.346105"} |