ppo-LunarLander-v2-unit1 / results.json
dcduplooy's picture
Initial commit of PPO model for LunarLander-v2
764d998
{"mean_reward": 261.11537452572117, "std_reward": 8.195681320288779, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-03-15T20:14:38.728027"}