hpourmodheji
commited on
Commit
•
3524a2a
1
Parent(s):
213d310
Upload PPO LunarLander-v2 trained agent
Browse files- README.md +1 -1
- config.json +1 -1
- ppo-LunarLander-v2.zip +2 -2
- ppo-LunarLander-v2/data +16 -16
- ppo-LunarLander-v2/policy.optimizer.pth +1 -1
- ppo-LunarLander-v2/policy.pth +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
README.md
CHANGED
@@ -16,7 +16,7 @@ model-index:
|
|
16 |
type: LunarLander-v2
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
-
value:
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
|
|
16 |
type: LunarLander-v2
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
+
value: 265.78 +/- 21.54
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7c4375b41ea0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7c4375b41f30>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7c4375b41fc0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7c4375b42050>", "_build": "<function ActorCriticPolicy._build at 0x7c4375b420e0>", "forward": "<function ActorCriticPolicy.forward at 0x7c4375b42170>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7c4375b42200>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7c4375b42290>", "_predict": "<function ActorCriticPolicy._predict at 0x7c4375b42320>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7c4375b423b0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7c4375b42440>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7c4375b424d0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7c4375b44040>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1700164406545607376, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGZhE77dfKc/s8GgvpXQ1L764Ua+VK+xvQAAAAAAAAAAmkFVPTr7az9mXy+6YPi1vgvh6zwwGp29AAAAAAAAAADNpRi9KSAwutzDQbZymISx/flkuCVnZTUAAIA/AACAP5Zxlz6ZcWE/6iLYPTWcv74xHGo+/uZRvAAAAAAAAAAARu0lPvHGxD41d6u+FzChvsAv1bwsP4y9AAAAAAAAAABNSJW9iAi3vEb+sT2WsAy9aeAXvjMqfL4AAIA/AACAPxq5gT0wDKI+yOBmvYG5Yb77HE49yRW4vAAAAAAAAAAAZtrnO1dKCD6eYba8hAI7vsaUqDyxCUe9AAAAAAAAAAAz0xO6wy0duopTbzoaBY81QBSpO1kRjrkAAIA/AACAPyBwTT4M/ro+4GmTvjGQt74qrre8M+ekvQAAAAAAAAAAzZxluy38lT65SYe9hsSIvm8GGb3D2Rk9AAAAAAAAAACazU89/snKPRG7Db4+aFa+3+8iPBIS97wAAAAAAAAAAM1EBb7krBI/770qPj0Xmr5F7Aw82p0cvAAAAAAAAAAA8wXSPa7sIT8+Lae9xXuyvvmlpDxVQQ+9AAAAAAAAAAANCK29pVrsPngKTTzl7KW+ms2DOWaJsb0AAAAAAAAAAACN6rwpEGe6MvaItiK1ALFlSQY5LT2lNQAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVPAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHIMnT3IuGuMAWyUTS0BjAF0lEdAmXOXtjTa03V9lChoBkdAcn5gwXZXdWgHTfQBaAhHQJl08v8IiTt1fZQoaAZHQHKvuYQarFRoB03WAWgIR0CZdVdxyXD4dX2UKGgGR0ByxfronrpraAdNiwFoCEdAmXVrvXsgMnV9lChoBkdAcGjmmLtNSWgHTTkBaAhHQJl1iM72crl1fZQoaAZHQHA1rngYP5JoB00AAWgIR0CZdxKkVN5/dX2UKGgGR0By5sQz1sciaAdNIwFoCEdAmXfw79ycTnV9lChoBkdAcKQ/47A+IWgHTV8BaAhHQJl4dstTUAl1fZQoaAZHQHKKLobGWD9oB00tAWgIR0CZeJoZAIIGdX2UKGgGR0ByG9qveP7vaAdNHAFoCEdAmXnVw5vLo3V9lChoBkdAcL9L2HtWuGgHTYIBaAhHQJl55mlImPZ1fZQoaAZHQG5snzpX6qNoB01WAWgIR0CZejNDMNc4dX2UKGgGR0BvcgXsPatcaAdNEAJoCEdAmXux+OOsDHV9lChoBkdAcahYcvM8o2gHTXQBaAhHQJl8rTy8SPF1fZQoaAZHQG45s3qAz55oB02jAWgIR0CZfNiVSn+AdX2UKGgGR0BygYIrvsqsaAdNGgFoCEdAmX08a0hNd3V9lChoBkdAcqI2lEZzgmgHTREBaAhHQJl+ZAu7HyV1fZQoaAZHQHBdANgBtDVoB00MAWgIR0CZf8k7wKBvdX2UKGgGR0Bv+VfPX05EaAdNSQFoCEdAmX/TwDvE0nV9lChoBkdAblITvAoG6mgHTSYCaAhHQJmAI6eXiR51fZQoaAZHQG9bzv7WNFVoB01PAWgIR0CZgHm5UcXFdX2UKGgGR0BvY3wmVqveaAdNYgFoCEdAmYDl/H5rQHV9lChoBkdAcVV4+8oQWmgHTRIBaAhHQJmBJpN9H+Z1fZQoaAZHQG63DzZpSJloB00SAWgIR0CZgUFLFn7IdX2UKGgGR0BwcVOEdvKmaAdNEwFoCEdAmYI9Vea8YnV9lChoBkdAbgi0lZ5iVmgHTRQBaAhHQJmCUBjnV5N1fZQoaAZHQG9wzjm0VrRoB01FAWgIR0CZhAcrAgxKdX2UKGgGR0BzGuQPqcEvaAdNLwFoCEdAmYTNFSbYsnV9lChoBkdAcN7Rvm5lOGgHTQwBaAhHQJmE24z7/GV1fZQoaAZHQHECa8cuJ1toB00kAWgIR0CZhWr7wazedX2UKGgGR0BuRulyimEXaAdNzwFoCEdAmYZaCHymRHV9lChoBkdAbQk5/b0voWgHTWMBaAhHQJmIM95hScd1fZQoaAZHQHDxAd8zAN5oB0v9aAhHQJmIRiKBNEh1fZQoaAZHQHFJttygf2doB00eAWgIR0CZiKk7OmiydX2UKGgGR0BzlxYMfA9FaAdNJQFoCEdAmYjXG8274HV9lChoBkdAcs45/b0voWgHTR4BaAhHQJmI9yvLX+V1fZQoaAZHQG9xAieNDMNoB0vyaAhHQJmJ67oSteV1fZQoaAZHQHJCvSYw7DFoB01CAWgIR0CZixTLGJemdX2UKGgGR0ByYQoXsPataAdNUwFoCEdAmYxKcI7eVXV9lChoBkdAcY1/XoTwlWgHTXIBaAhHQJmNJe5WilB1fZQoaAZHQHKp6mwaBI5oB01LAWgIR0CZjTOdXko4dX2UKGgGR0BuBHmV7hNuaAdN2gFoCEdAmY2v5HmRvHV9lChoBkdAcfw5NGmUGGgHTQwBaAhHQJmN8STQmeF1fZQoaAZHQHI6QzLwF1VoB0v5aAhHQJmN+GfwqiJ1fZQoaAZHQG61+3Ytg8doB00NAWgIR0CZjgMsYl6adX2UKGgGR0BwE2tbLU1AaAdNHgFoCEdAmZAgSOBDonV9lChoBkdAcSgqS5iEx2gHTSEBaAhHQJmTk2eg+Ql1fZQoaAZHQHBUTqW1MM9oB01PAWgIR0CZlW6Rhc7hdX2UKGgGR0BxgbfNzKcNaAdNPwFoCEdAmZWHPZ7HAHV9lChoBkdAclBm9g4OtmgHTU0BaAhHQJmWb3xnWat1fZQoaAZHQHDaLAckt29oB01nAWgIR0CZloyHVPN3dX2UKGgGR0BydwLG7z06aAdNBwFoCEdAmZdhtUGVzXV9lChoBkdAcBFB7/n4f2gHTUgBaAhHQJmvlQ+EAYJ1fZQoaAZHQHHPy3b212JoB00UAWgIR0CZsKRHPNVzdX2UKGgGR0ByjONGViWnaAdNPQFoCEdAmbISp3os7XV9lChoBkdAcAc2rGR3eWgHTTIBaAhHQJmyMV45cTt1fZQoaAZHQHHXhQemvW9oB01JAWgIR0CZsy0cfeUIdX2UKGgGR0Bw3QgB91EFaAdNTAFoCEdAmbNPuw5eaHV9lChoBkdAcjMXKKYRd2gHTYoBaAhHQJm1RnezlcR1fZQoaAZHQHDUNlVcUudoB00HAWgIR0CZt0xwAEMcdX2UKGgGR0Bv4Lbg0j1PaAdNKAJoCEdAmbfusPrfL3V9lChoBkdAcY5JVbRne2gHTVgBaAhHQJm4TmeUY9B1fZQoaAZHQHGh3hXKbKBoB01HAWgIR0CZuPQGwA2idX2UKGgGR0ByQMb6xgRcaAdNWwFoCEdAmbmMu3+db3V9lChoBkdAcTphmXgLqmgHTT0BaAhHQJm548hcJMR1fZQoaAZHQHF38uJ1q35oB01CAWgIR0CZu15BTn7pdX2UKGgGR0BxbZ/Aj6eoaAdNiAFoCEdAmbvB5kbxVnV9lChoBkdAcM0KNhmXgWgHTSwBaAhHQJm8CnHeaa11fZQoaAZHQG4aqVQhwERoB02AAWgIR0CZvFposZpBdX2UKGgGR0BywrHT7VJ+aAdNQwFoCEdAmb3nN9ph4XV9lChoBkdAcnk/336AOWgHTQsBaAhHQJm+RKVY6n11fZQoaAZHQHBQ2I42jwhoB01wAWgIR0CZvlPkq+ajdX2UKGgGR0Bc3a1og3cYaAdN6ANoCEdAmb+O89Oh03V9lChoBkdAccbKmKqGUWgHTSMBaAhHQJnAue5Fw1l1fZQoaAZHQHIAvFNtZV5oB00LAWgIR0CZwXGIsRQKdX2UKGgGR0Bxr9sYVIqcaAdN4wJoCEdAmcK7i6xxDXV9lChoBkdAc3HdyT6i02gHTVcBaAhHQJnDDaxoqTd1fZQoaAZHQHGGvffoA4poB0v2aAhHQJnDPechC+l1fZQoaAZHQG9XM5wOvuBoB01mAWgIR0CZw+hC+lCUdX2UKGgGR0BwfjeN1hb4aAdNOQFoCEdAmcP7kXDWLHV9lChoBkdAcHhGOMl1KWgHTf4BaAhHQJnEDLeQ+2V1fZQoaAZHQHFOML0Bfa9oB01OAWgIR0CZxEVeruIAdX2UKGgGR0Bw8E6RyOrAaAdNEwFoCEdAmcSIwh4dIXV9lChoBkdAcfd9gWrOq2gHTQ0BaAhHQJnGR5rxiG51fZQoaAZHQHKCKTnq3VloB01PAWgIR0CZxsha1TisdX2UKGgGR0BwxCpvP1L8aAdL9GgIR0CZxzTmnwXqdX2UKGgGR0Bw7De54GD+aAdNiwFoCEdAmchoO6NEPXV9lChoBkdAcOC+kP+XJGgHTW4BaAhHQJnJ1zFMqSZ1fZQoaAZHQDXDugHu7YloB0vGaAhHQJnKFZuAI6d1fZQoaAZHQHFWpnL7oB9oB0vjaAhHQJnK/6Fdszl1fZQoaAZHQHNR/G6wt8NoB01FAWgIR0CZyyM1jy4GdX2UKGgGR0Bw8p05lvqDaAdNnwFoCEdAmctudf9gnnV9lChoBkdAcIqQoCuEEmgHTTgBaAhHQJnLbb48EFJ1fZQoaAZHQHILyRB/qgRoB00yAWgIR0CZzKPldTo/dX2UKGgGR0Bybup2ll9SaAdNDAFoCEdAmcylZcLSeHV9lChoBkdAcHUDjR2KVWgHTQYBaAhHQJnMuPyTY/V1fZQoaAZHQHGSAXhwVCZoB01RAWgIR0CZzTZvUBn0dX2UKGgGR0BvMk8mrsByaAdNUQFoCEdAmc2alP8AJnV9lChoBkdAcmxZNwiqyWgHTQABaAhHQJnOH+Idlup1fZQoaAZHQHBDLkbPyCpoB0v0aAhHQJnOlaEBbOh1ZS4="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-5.15.120+-x86_64-with-glibc2.35 # 1 SMP Wed Aug 30 11:19:59 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu118", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x788cbcef5750>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x788cbcef57e0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x788cbcef5870>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x788cbcef5900>", "_build": "<function ActorCriticPolicy._build at 0x788cbcef5990>", "forward": "<function ActorCriticPolicy.forward at 0x788cbcef5a20>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x788cbcef5ab0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x788cbcef5b40>", "_predict": "<function ActorCriticPolicy._predict at 0x788cbcef5bd0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x788cbcef5c60>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x788cbcef5cf0>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x788cbcef5d80>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x788cbce8adc0>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1700534278387440457, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAADMg1Txc+1K66M1WuN7KOrNzxaM5a1d9NwAAgD8AAIA/ZquDvPaUMrpNqNU7izEBOIIxEzoQVCU2AACAPwAAgD+auZm7KRw6uFG3QznN+Q806UBYO/XhZrgAAIA/AACAP9r5BD6wNGc/XSHevQp7sb7t2+Y9DH8bvgAAAAAAAAAAM16MvIVby7lVfFU5vI2jsypUJruVuHu4AACAPwAAgD/Ndem8SN+PugqzVTtqQiE45uSeOTf0BroAAIA/AACAP83v8j11PIw/XgIYPp6S2b5yLRQ+6+0ovQAAAAAAAAAAgL1Rva6HjLrTp4c72mgxOJl/NzuGjzO4AACAPwAAgD8Ack+89phAum3/YbrJFg22hh/zuo3AhDkAAIA/AACAP01As72PgkK6QoqhuqfvdbaJ9XM7YFzlNQAAAAAAAIA/ALonPOFspbouHp047183tpY5SjqU7rG3AACAPwAAgD9zY5U99ow9uiytB7pMNC4zPqxZO/qoSrMAAIA/AACAP5pF7Tv4hKk8KjaYvbFh6L0DnqW8Vqm6ugAAAAAAAAAAGtrsvYWgqz6oe+q82TqRvimuAr5k+xy9AAAAAAAAAADA6Zo94kdoP+oJWj31kbq+8RK8PTOAnLwAAAAAAAAAAJr9JbwpdEu6LsK0t96wnbJO1gS7EJnVNgAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVQAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQFq+MNtqHoKMAWyUTegDjAF0lEdAk8yf/JeVs3V9lChoBkdAZsTiVB2OhmgHTegDaAhHQJPOcE6kqMF1fZQoaAZHQGQDemNzbN9oB03oA2gIR0CT0EZR8+ibdX2UKGgGR0Bf2GBvrGBGaAdN6ANoCEdAk9lGBJ7LMnV9lChoBkdAZ0OeDFqBVmgHTegDaAhHQJPokRL9MsZ1fZQoaAZHQGLScdPtUn5oB03oA2gIR0CT7GujRD1HdX2UKGgGR0Bn2xA0Kqn4aAdN6ANoCEdAk+z/91loUXV9lChoBkdAXyb5GjKxLWgHTegDaAhHQJPzBKbrkbR1fZQoaAZHQFFduSOinHhoB0vCaAhHQJP1SbZvkzZ1fZQoaAZHQGRKTguRLbpoB03oA2gIR0CT90LIxQBQdX2UKGgGR0BhfI7V8Ti9aAdN6ANoCEdAk/eswco6S3V9lChoBkdAaLrC53C9AWgHTegDaAhHQJP6gWWQfZF1fZQoaAZHQGH2X2mHgxdoB03oA2gIR0CT/TrNnoPkdX2UKGgGR0BgqhEv0yxiaAdN6ANoCEdAk/7cFhXr+3V9lChoBkdAZBujFAE+xGgHTegDaAhHQJQlSnk1dgR1fZQoaAZHQGRGZc1O0sxoB03oA2gIR0CUJ9bB42S/dX2UKGgGR0BlkexMWXTmaAdN6ANoCEdAlCmeHaewtHV9lChoBkdAY/NnB+F10WgHTegDaAhHQJQs64b0e2d1fZQoaAZHQGNOMVk+X7doB03oA2gIR0CULyZAprk9dX2UKGgGR0BnLEse4kNXaAdN6ANoCEdAlDB7Zi/fwnV9lChoBkdAZfyXizcAR2gHTegDaAhHQJQ22Nm16Vt1fZQoaAZHQGgc0sOG0u1oB03oA2gIR0CURJz2OAAidX2UKGgGR0Bm/Ap6QeV+aAdN6ANoCEdAlEUfY4ACGXV9lChoBkdAZGsUrTYukGgHTegDaAhHQJRKfwUg0TF1fZQoaAZHQGZA+ws5GSZoB03oA2gIR0CUTKxhlUZOdX2UKGgGR0Bo++0CzTnaaAdN6ANoCEdAlE5606YE4nV9lChoBkdAY/n433pOe2gHTegDaAhHQJRO4Ds+mnB1fZQoaAZHQGYCMWGh24doB03oA2gIR0CUUadzXBgvdX2UKGgGR0BkresgdOqOaAdN6ANoCEdAlFQo0EX+EXV9lChoBkdAZ2Nn0TURWmgHTegDaAhHQJRVRDTjNpx1fZQoaAZHQGaPidBjWkJoB03oA2gIR0CUds6Zpi7TdX2UKGgGR0Bjl7sByS3caAdN6ANoCEdAlHifpyIYWXV9lChoBkdAZpX1XeWOZWgHTegDaAhHQJR5zqJMxoJ1fZQoaAZHQEDao2n889xoB0vraAhHQJR6Ay44Ia91fZQoaAZHQGBvGFzuF6BoB03oA2gIR0CUe+LofSx8dX2UKGgGR0BksUjPfKp2aAdN6ANoCEdAlH07l/6O53V9lChoBkdAZL9vlU6xPmgHTegDaAhHQJR+ZxPwd811fZQoaAZHQGMQntWuHN5oB03oA2gIR0CUhHk/8l5XdX2UKGgGR0Ah/ltj0+TvaAdL32gIR0CUkT+wTufFdX2UKGgGR0BnBQegctGvaAdN6ANoCEdAlJIdJe3QU3V9lChoBkdAY05MSsbNr2gHTegDaAhHQJSSoNKAavR1fZQoaAZHQGSfs189fTloB03oA2gIR0CUmAr1dxACdX2UKGgGR0BoGIOe8PFvaAdN6ANoCEdAlJrPz8P4EnV9lChoBkdAaDhNpudf9mgHTegDaAhHQJSdZyU9pyp1fZQoaAZHQGOKJsoDxLFoB03oA2gIR0CUnfAk9lmOdX2UKGgGR0BgNsdJaq0daAdN6ANoCEdAlKGZL/S6UnV9lChoBkdAY1J4A0bcXWgHTegDaAhHQJSlDfwZwXJ1fZQoaAZHQG6+SfcvduZoB021A2gIR0CUwoJY1YQrdX2UKGgGR0BiBgM2FWXDaAdN6ANoCEdAlMcogNgBtHV9lChoBkdAY324FzMibGgHTegDaAhHQJTIVbpu/Dd1fZQoaAZHQGT2Dzyz5XVoB03oA2gIR0CUyIr92ovSdX2UKGgGR0BjNxHI6r/9aAdN6ANoCEdAlMp0GzKLbnV9lChoBkdAZjPsi0OVgWgHTegDaAhHQJTL3p7kXDZ1fZQoaAZHQGbHjx9XtBxoB03oA2gIR0CUzSKcNH6NdX2UKGgGR0BhUM9Mbm2caAdN6ANoCEdAlOHY+8oQWnV9lChoBkdAZaH/R3NcGGgHTegDaAhHQJTi3/7zkIZ1fZQoaAZHQGSrcMEzO5doB03oA2gIR0CU44A9FF2FdX2UKGgGR0BjnWWnjyWiaAdN6ANoCEdAlOoEFfReC3V9lChoBkdAZABrsSkCWGgHTegDaAhHQJTs27kGRmt1fZQoaAZHQGdPFHBk7OpoB03oA2gIR0CU7yVJcxCZdX2UKGgGR0BgBF5dGAkLaAdN6ANoCEdAlO+mBvrGBHV9lChoBkdAYJ6FUQ04zmgHTegDaAhHQJTzQjLSuyN1fZQoaAZHQGJjwWvbGm1oB03oA2gIR0CU9qytV7x/dX2UKGgGR0BwQuHck+otaAdN5QJoCEdAlPngUtZmqnV9lChoBkdAY8hqXWvr4WgHTegDaAhHQJT+p7HAAQx1fZQoaAZHQGRr07bL2YhoB03oA2gIR0CVFF8O09hadX2UKGgGR0BngHXwsoUjaAdN6ANoCEdAlRWf2PDHfnV9lChoBkdAZczWxyGSIWgHTegDaAhHQJUV2rT6SDB1fZQoaAZHQGFZHhCMPz5oB03oA2gIR0CVF9jcVQANdX2UKGgGR0Bmp0mfGuLaaAdN6ANoCEdAlRrUMoc7yXV9lChoBkdAbD2Hpr1ui2gHTSgCaAhHQJUkrkGRmsh1fZQoaAZHQGGqgxBVuJloB03oA2gIR0CVNPqrBCUpdX2UKGgGR0BiEw4XGff5aAdN6ANoCEdAlTYgrH2h7HV9lChoBkdAYYr0wrUb1mgHTegDaAhHQJU2zRa5f+l1fZQoaAZHQGTqzTF2mpFoB03oA2gIR0CVPCj+JgstdX2UKGgGR0BgzCi0v4/NaAdN6ANoCEdAlT4IKtxMnXV9lChoBkdAb+3zAeq7y2gHTWoBaAhHQJU+bVtoBaN1fZQoaAZHQGSI5qubI91oB03oA2gIR0CVP4bcoH9ndX2UKGgGR0BgXfA9FF2FaAdN6ANoCEdAlT/hS9/SY3V9lChoBkdAYdJRFZxJd2gHTegDaAhHQJVEy9SMtK91fZQoaAZHQF4rAy2x6fJoB03oA2gIR0CVR/YHxBmgdX2UKGgGR0BRRoSDh99daAdL7WgIR0CVS3PRRdhRdX2UKGgGR0Bo72h/RVp9aAdN6ANoCEdAlUxXMhX8wnV9lChoBkdAZuWRVZLZjGgHTegDaAhHQJViIRChN/R1fZQoaAZHQGcIkSVW0Z5oB03oA2gIR0CVY706YE4edX2UKGgGR0BoRTcynDR/aAdN6ANoCEdAlWP/mozeoHV9lChoBkdAYY+1k1/DtWgHTegDaAhHQJVmW3I+4b11fZQoaAZHQGiVngP3BYVoB03oA2gIR0CVadpUxVQzdX2UKGgGR0BupSr7wazeaAdNLQNoCEdAlYBQWrOqvXV9lChoBkdAYT+Iv8IiT2gHTegDaAhHQJWCLOryUcJ1fZQoaAZHQGRxS925hBtoB03oA2gIR0CVguNe+mFbdX2UKGgGR0BkZerELpiaaAdN6ANoCEdAlYNaHbh3q3V9lChoBkdAYWKNlRP422gHTegDaAhHQJWHrlRxcVx1fZQoaAZHQF3iLwF1SwZoB03oA2gIR0CVib7Wd3B6dX2UKGgGR0Bj4rsMRYigaAdN6ANoCEdAlYrGDUVi4XV9lChoBkdAZqb0cwQDm2gHTegDaAhHQJWP0fp2U0N1fZQoaAZHQGa7eBYmsvJoB03oA2gIR0CVkvmlZX+3dX2UKGgGR0BomxDu0CzUaAdN6ANoCEdAlZZ8xfv4NHV9lChoBkdAZNOjWTX8O2gHTegDaAhHQJWXZ0/4Zdh1fZQoaAZHQHKVkhmoR7JoB00mAWgIR0CVl+GT9sJqdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-5.15.120+-x86_64-with-glibc2.35 # 1 SMP Wed Aug 30 11:19:59 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu118", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
ppo-LunarLander-v2.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3070d96ab81b1eb873d7a9ad1a718cbf0074aca0d55172a4c2b6f3f1af338f5
|
3 |
+
size 148050
|
ppo-LunarLander-v2/data
CHANGED
@@ -4,20 +4,20 @@
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
11 |
-
"_build": "<function ActorCriticPolicy._build at
|
12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
-
"_abc_impl": "<_abc._abc_data object at
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {},
|
@@ -26,12 +26,12 @@
|
|
26 |
"_num_timesteps_at_start": 0,
|
27 |
"seed": null,
|
28 |
"action_noise": null,
|
29 |
-
"start_time":
|
30 |
"learning_rate": 0.0003,
|
31 |
"tensorboard_log": null,
|
32 |
"_last_obs": {
|
33 |
":type:": "<class 'numpy.ndarray'>",
|
34 |
-
":serialized:": "
|
35 |
},
|
36 |
"_last_episode_starts": {
|
37 |
":type:": "<class 'numpy.ndarray'>",
|
@@ -45,7 +45,7 @@
|
|
45 |
"_stats_window_size": 100,
|
46 |
"ep_info_buffer": {
|
47 |
":type:": "<class 'collections.deque'>",
|
48 |
-
":serialized:": "
|
49 |
},
|
50 |
"ep_success_buffer": {
|
51 |
":type:": "<class 'collections.deque'>",
|
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x788cbcef5750>",
|
8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x788cbcef57e0>",
|
9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x788cbcef5870>",
|
10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x788cbcef5900>",
|
11 |
+
"_build": "<function ActorCriticPolicy._build at 0x788cbcef5990>",
|
12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x788cbcef5a20>",
|
13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x788cbcef5ab0>",
|
14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x788cbcef5b40>",
|
15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x788cbcef5bd0>",
|
16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x788cbcef5c60>",
|
17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x788cbcef5cf0>",
|
18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x788cbcef5d80>",
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
+
"_abc_impl": "<_abc._abc_data object at 0x788cbce8adc0>"
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {},
|
|
|
26 |
"_num_timesteps_at_start": 0,
|
27 |
"seed": null,
|
28 |
"action_noise": null,
|
29 |
+
"start_time": 1700534278387440457,
|
30 |
"learning_rate": 0.0003,
|
31 |
"tensorboard_log": null,
|
32 |
"_last_obs": {
|
33 |
":type:": "<class 'numpy.ndarray'>",
|
34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAADMg1Txc+1K66M1WuN7KOrNzxaM5a1d9NwAAgD8AAIA/ZquDvPaUMrpNqNU7izEBOIIxEzoQVCU2AACAPwAAgD+auZm7KRw6uFG3QznN+Q806UBYO/XhZrgAAIA/AACAP9r5BD6wNGc/XSHevQp7sb7t2+Y9DH8bvgAAAAAAAAAAM16MvIVby7lVfFU5vI2jsypUJruVuHu4AACAPwAAgD/Ndem8SN+PugqzVTtqQiE45uSeOTf0BroAAIA/AACAP83v8j11PIw/XgIYPp6S2b5yLRQ+6+0ovQAAAAAAAAAAgL1Rva6HjLrTp4c72mgxOJl/NzuGjzO4AACAPwAAgD8Ack+89phAum3/YbrJFg22hh/zuo3AhDkAAIA/AACAP01As72PgkK6QoqhuqfvdbaJ9XM7YFzlNQAAAAAAAIA/ALonPOFspbouHp047183tpY5SjqU7rG3AACAPwAAgD9zY5U99ow9uiytB7pMNC4zPqxZO/qoSrMAAIA/AACAP5pF7Tv4hKk8KjaYvbFh6L0DnqW8Vqm6ugAAAAAAAAAAGtrsvYWgqz6oe+q82TqRvimuAr5k+xy9AAAAAAAAAADA6Zo94kdoP+oJWj31kbq+8RK8PTOAnLwAAAAAAAAAAJr9JbwpdEu6LsK0t96wnbJO1gS7EJnVNgAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
35 |
},
|
36 |
"_last_episode_starts": {
|
37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
45 |
"_stats_window_size": 100,
|
46 |
"ep_info_buffer": {
|
47 |
":type:": "<class 'collections.deque'>",
|
48 |
+
":serialized:": "gAWVQAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQFq+MNtqHoKMAWyUTegDjAF0lEdAk8yf/JeVs3V9lChoBkdAZsTiVB2OhmgHTegDaAhHQJPOcE6kqMF1fZQoaAZHQGQDemNzbN9oB03oA2gIR0CT0EZR8+ibdX2UKGgGR0Bf2GBvrGBGaAdN6ANoCEdAk9lGBJ7LMnV9lChoBkdAZ0OeDFqBVmgHTegDaAhHQJPokRL9MsZ1fZQoaAZHQGLScdPtUn5oB03oA2gIR0CT7GujRD1HdX2UKGgGR0Bn2xA0Kqn4aAdN6ANoCEdAk+z/91loUXV9lChoBkdAXyb5GjKxLWgHTegDaAhHQJPzBKbrkbR1fZQoaAZHQFFduSOinHhoB0vCaAhHQJP1SbZvkzZ1fZQoaAZHQGRKTguRLbpoB03oA2gIR0CT90LIxQBQdX2UKGgGR0BhfI7V8Ti9aAdN6ANoCEdAk/eswco6S3V9lChoBkdAaLrC53C9AWgHTegDaAhHQJP6gWWQfZF1fZQoaAZHQGH2X2mHgxdoB03oA2gIR0CT/TrNnoPkdX2UKGgGR0BgqhEv0yxiaAdN6ANoCEdAk/7cFhXr+3V9lChoBkdAZBujFAE+xGgHTegDaAhHQJQlSnk1dgR1fZQoaAZHQGRGZc1O0sxoB03oA2gIR0CUJ9bB42S/dX2UKGgGR0BlkexMWXTmaAdN6ANoCEdAlCmeHaewtHV9lChoBkdAY/NnB+F10WgHTegDaAhHQJQs64b0e2d1fZQoaAZHQGNOMVk+X7doB03oA2gIR0CULyZAprk9dX2UKGgGR0BnLEse4kNXaAdN6ANoCEdAlDB7Zi/fwnV9lChoBkdAZfyXizcAR2gHTegDaAhHQJQ22Nm16Vt1fZQoaAZHQGgc0sOG0u1oB03oA2gIR0CURJz2OAAidX2UKGgGR0Bm/Ap6QeV+aAdN6ANoCEdAlEUfY4ACGXV9lChoBkdAZGsUrTYukGgHTegDaAhHQJRKfwUg0TF1fZQoaAZHQGZA+ws5GSZoB03oA2gIR0CUTKxhlUZOdX2UKGgGR0Bo++0CzTnaaAdN6ANoCEdAlE5606YE4nV9lChoBkdAY/n433pOe2gHTegDaAhHQJRO4Ds+mnB1fZQoaAZHQGYCMWGh24doB03oA2gIR0CUUadzXBgvdX2UKGgGR0BkresgdOqOaAdN6ANoCEdAlFQo0EX+EXV9lChoBkdAZ2Nn0TURWmgHTegDaAhHQJRVRDTjNpx1fZQoaAZHQGaPidBjWkJoB03oA2gIR0CUds6Zpi7TdX2UKGgGR0Bjl7sByS3caAdN6ANoCEdAlHifpyIYWXV9lChoBkdAZpX1XeWOZWgHTegDaAhHQJR5zqJMxoJ1fZQoaAZHQEDao2n889xoB0vraAhHQJR6Ay44Ia91fZQoaAZHQGBvGFzuF6BoB03oA2gIR0CUe+LofSx8dX2UKGgGR0BksUjPfKp2aAdN6ANoCEdAlH07l/6O53V9lChoBkdAZL9vlU6xPmgHTegDaAhHQJR+ZxPwd811fZQoaAZHQGMQntWuHN5oB03oA2gIR0CUhHk/8l5XdX2UKGgGR0Ah/ltj0+TvaAdL32gIR0CUkT+wTufFdX2UKGgGR0BnBQegctGvaAdN6ANoCEdAlJIdJe3QU3V9lChoBkdAY05MSsbNr2gHTegDaAhHQJSSoNKAavR1fZQoaAZHQGSfs189fTloB03oA2gIR0CUmAr1dxACdX2UKGgGR0BoGIOe8PFvaAdN6ANoCEdAlJrPz8P4EnV9lChoBkdAaDhNpudf9mgHTegDaAhHQJSdZyU9pyp1fZQoaAZHQGOKJsoDxLFoB03oA2gIR0CUnfAk9lmOdX2UKGgGR0BgNsdJaq0daAdN6ANoCEdAlKGZL/S6UnV9lChoBkdAY1J4A0bcXWgHTegDaAhHQJSlDfwZwXJ1fZQoaAZHQG6+SfcvduZoB021A2gIR0CUwoJY1YQrdX2UKGgGR0BiBgM2FWXDaAdN6ANoCEdAlMcogNgBtHV9lChoBkdAY324FzMibGgHTegDaAhHQJTIVbpu/Dd1fZQoaAZHQGT2Dzyz5XVoB03oA2gIR0CUyIr92ovSdX2UKGgGR0BjNxHI6r/9aAdN6ANoCEdAlMp0GzKLbnV9lChoBkdAZjPsi0OVgWgHTegDaAhHQJTL3p7kXDZ1fZQoaAZHQGbHjx9XtBxoB03oA2gIR0CUzSKcNH6NdX2UKGgGR0BhUM9Mbm2caAdN6ANoCEdAlOHY+8oQWnV9lChoBkdAZaH/R3NcGGgHTegDaAhHQJTi3/7zkIZ1fZQoaAZHQGSrcMEzO5doB03oA2gIR0CU44A9FF2FdX2UKGgGR0BjnWWnjyWiaAdN6ANoCEdAlOoEFfReC3V9lChoBkdAZABrsSkCWGgHTegDaAhHQJTs27kGRmt1fZQoaAZHQGdPFHBk7OpoB03oA2gIR0CU7yVJcxCZdX2UKGgGR0BgBF5dGAkLaAdN6ANoCEdAlO+mBvrGBHV9lChoBkdAYJ6FUQ04zmgHTegDaAhHQJTzQjLSuyN1fZQoaAZHQGJjwWvbGm1oB03oA2gIR0CU9qytV7x/dX2UKGgGR0BwQuHck+otaAdN5QJoCEdAlPngUtZmqnV9lChoBkdAY8hqXWvr4WgHTegDaAhHQJT+p7HAAQx1fZQoaAZHQGRr07bL2YhoB03oA2gIR0CVFF8O09hadX2UKGgGR0BngHXwsoUjaAdN6ANoCEdAlRWf2PDHfnV9lChoBkdAZczWxyGSIWgHTegDaAhHQJUV2rT6SDB1fZQoaAZHQGFZHhCMPz5oB03oA2gIR0CVF9jcVQANdX2UKGgGR0Bmp0mfGuLaaAdN6ANoCEdAlRrUMoc7yXV9lChoBkdAbD2Hpr1ui2gHTSgCaAhHQJUkrkGRmsh1fZQoaAZHQGGqgxBVuJloB03oA2gIR0CVNPqrBCUpdX2UKGgGR0BiEw4XGff5aAdN6ANoCEdAlTYgrH2h7HV9lChoBkdAYYr0wrUb1mgHTegDaAhHQJU2zRa5f+l1fZQoaAZHQGTqzTF2mpFoB03oA2gIR0CVPCj+JgstdX2UKGgGR0BgzCi0v4/NaAdN6ANoCEdAlT4IKtxMnXV9lChoBkdAb+3zAeq7y2gHTWoBaAhHQJU+bVtoBaN1fZQoaAZHQGSI5qubI91oB03oA2gIR0CVP4bcoH9ndX2UKGgGR0BgXfA9FF2FaAdN6ANoCEdAlT/hS9/SY3V9lChoBkdAYdJRFZxJd2gHTegDaAhHQJVEy9SMtK91fZQoaAZHQF4rAy2x6fJoB03oA2gIR0CVR/YHxBmgdX2UKGgGR0BRRoSDh99daAdL7WgIR0CVS3PRRdhRdX2UKGgGR0Bo72h/RVp9aAdN6ANoCEdAlUxXMhX8wnV9lChoBkdAZuWRVZLZjGgHTegDaAhHQJViIRChN/R1fZQoaAZHQGcIkSVW0Z5oB03oA2gIR0CVY706YE4edX2UKGgGR0BoRTcynDR/aAdN6ANoCEdAlWP/mozeoHV9lChoBkdAYY+1k1/DtWgHTegDaAhHQJVmW3I+4b11fZQoaAZHQGiVngP3BYVoB03oA2gIR0CVadpUxVQzdX2UKGgGR0BupSr7wazeaAdNLQNoCEdAlYBQWrOqvXV9lChoBkdAYT+Iv8IiT2gHTegDaAhHQJWCLOryUcJ1fZQoaAZHQGRxS925hBtoB03oA2gIR0CVguNe+mFbdX2UKGgGR0BkZerELpiaaAdN6ANoCEdAlYNaHbh3q3V9lChoBkdAYWKNlRP422gHTegDaAhHQJWHrlRxcVx1fZQoaAZHQF3iLwF1SwZoB03oA2gIR0CVib7Wd3B6dX2UKGgGR0Bj4rsMRYigaAdN6ANoCEdAlYrGDUVi4XV9lChoBkdAZqb0cwQDm2gHTegDaAhHQJWP0fp2U0N1fZQoaAZHQGa7eBYmsvJoB03oA2gIR0CVkvmlZX+3dX2UKGgGR0BomxDu0CzUaAdN6ANoCEdAlZZ8xfv4NHV9lChoBkdAZNOjWTX8O2gHTegDaAhHQJWXZ0/4Zdh1fZQoaAZHQHKVkhmoR7JoB00mAWgIR0CVl+GT9sJqdWUu"
|
49 |
},
|
50 |
"ep_success_buffer": {
|
51 |
":type:": "<class 'collections.deque'>",
|
ppo-LunarLander-v2/policy.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 88362
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ddf1d700e1d57763e6f92f12e1017e701135b5361a998fec31abde3a65d4393a
|
3 |
size 88362
|
ppo-LunarLander-v2/policy.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 43762
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4474dab367baf357d0d7156bf9e3426c0e904fd8ede1917835503fad515f6245
|
3 |
size 43762
|
replay.mp4
CHANGED
Binary files a/replay.mp4 and b/replay.mp4 differ
|
|
results.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"mean_reward":
|
|
|
1 |
+
{"mean_reward": 265.77730740000004, "std_reward": 21.543547160507583, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-11-21T03:05:44.705274"}
|