ppo-LunarLander-v2 / results.json
Michunie's picture
Second try - changes parameters: n_steps = 2048, batch_size = 128
6d06b7e
{"mean_reward": 284.2997220133261, "std_reward": 17.281453080910197, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-12-10T23:38:45.167616"}