Commit
•
4545000
1
Parent(s):
5b25308
Done
Browse files- README.md +1 -1
- config.json +1 -1
- ppo-LunarLander-v2.zip +2 -2
- ppo-LunarLander-v2/data +16 -16
- ppo-LunarLander-v2/policy.optimizer.pth +1 -1
- ppo-LunarLander-v2/policy.pth +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
README.md
CHANGED
@@ -16,7 +16,7 @@ model-index:
|
|
16 |
type: LunarLander-v2
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
-
value:
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
|
|
16 |
type: LunarLander-v2
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
+
value: 258.57 +/- 20.65
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x793c9e660790>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x793c9e660820>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x793c9e6608b0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x793c9e660940>", "_build": "<function ActorCriticPolicy._build at 0x793c9e6609d0>", "forward": "<function ActorCriticPolicy.forward at 0x793c9e660a60>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x793c9e660af0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x793c9e660b80>", "_predict": "<function ActorCriticPolicy._predict at 0x793c9e660c10>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x793c9e660ca0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x793c9e660d30>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x793c9e660dc0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x793c9e7f3480>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1710350435041850379, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGBHaz4WOM0+vsf8veY3VL5Vxc47kJlaPQAAAAAAAAAAk4MXvqSrL7sua0m6P/0Rt8pNPTyD3Wk5AACAPwAAgD9ajJ8+o6uoP3tsZz4wGx2+3SBSPkCBmr0AAAAAAAAAAGZmV7r2HEe6/RnKOw9mzjdFqyq67fQoNgAAgD8AAIA/QFjWPRT0jrolYB+8+Vblte7RS7l+elA1AAAAAAAAgD+aoiK+Q8ohvPRnyruJjP65XDaIPVB97zoAAIA/AACAPyZ1zD3gRJk/Xj4/Pj3wAr4CZ4s9rl1qPQAAAAAAAAAAM1GwPPZISLoKSwq7RarQNV/SJzt1cD61AACAPwAAgD9zrLa9j/5ZuicLqjt+Axs4muhxOrAYg7oAAAAAAACAP7P7pL4M6+s+i6LVPSq6gb4R+ci9kD+SPQAAAAAAAAAAZuIVvOEo6z7j4EE+MkQYvixkRT0opNs9AAAAAAAAAAANEu49BDHwPjFEir33y4C+YeFePR4rxzwAAAAAAAAAAJrrVD32g5g/CEjgPd/Sh76LwKA9CrCFvQAAAAAAAAAAAHSGu2Ygoj+bv3k8spq0vvhPT7xNIvu8AAAAAAAAAABNvN29ITDiPSLxBT63U0W+fwkCPE1vHzwAAAAAAAAAAEB5oz3D0Ta6ZptAOaAuqTTnvBk4yJ1duAAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVOwwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQGYFcfeUILSMAWyUTegDjAF0lEdAkhAqHfuTinV9lChoBkdAN6bBTGYKIGgHTQsBaAhHQJIQ0VGkN4J1fZQoaAZHQGNyVM23rlhoB03oA2gIR0CSE5CiRGMGdX2UKGgGR0BkeBtcfNiZaAdN6ANoCEdAkij/kmx+rnV9lChoBkdAYwhJp35eq2gHTegDaAhHQJIqQ+jdpIt1fZQoaAZHQFro0F8ohIRoB03oA2gIR0CSKo/zJ6ppdX2UKGgGR0Bc/n/1g6U8aAdN6ANoCEdAkitkXxe9jHV9lChoBkdAZO5HR1HOKWgHTegDaAhHQJI0HszEaVF1fZQoaAZHQGNpyzXz19RoB03oA2gIR0CSPPPu5SWJdX2UKGgGR0BgMOMqBmPHaAdN6ANoCEdAkkPSUPhAGHV9lChoBkdAX13DuSfUWmgHTegDaAhHQJJF27SRbKR1fZQoaAZHQGTmHh86V+toB03oA2gIR0CSSK5Qgs9TdX2UKGgGR0BmXmo3rD64aAdN6ANoCEdAkk7VS0jTrnV9lChoBkdAYYphrnDBM2gHTegDaAhHQJJXCpXIU8F1fZQoaAZHQGGsJyyUs4FoB03oA2gIR0CSXYuBczIndX2UKGgGR0BmbCLXL/0eaAdN6ANoCEdAkmASJ9AoonV9lChoBkdAYr9vXK8tgGgHTegDaAhHQJJjdKh+OOt1fZQoaAZHQGOb4FzMibFoB03oA2gIR0CSZGY64lQedX2UKGgGR0Bi4rqQiiZfaAdN6ANoCEdAkmgpJ5E+gXV9lChoBkdAYEu+dsi0OWgHTegDaAhHQJJ9cddVvMt1fZQoaAZHQGLb7FS88LdoB03oA2gIR0CSfpeLehwmdX2UKGgGR0Bl+cQmNR3vaAdN6ANoCEdAkn7iLhrFfnV9lChoBkdAZN4q4H5aeWgHTegDaAhHQJJ/4ZEUj9p1fZQoaAZHQGKL19fCyhVoB03oA2gIR0CSiD4YrJ8wdX2UKGgGR0BFx8Vgx8D0aAdL+mgIR0CSiGnm7rcCdX2UKGgGR0BQB79ETg2qaAdL7mgIR0CSjgnCO3lTdX2UKGgGR0Bmlak0rK/3aAdN6ANoCEdAkpDqv/zasnV9lChoBkdAYSu3DNyHVWgHTegDaAhHQJKZt/DtPYZ1fZQoaAZHQGSyeHzpX6toB03oA2gIR0CSm7Bdld1MdX2UKGgGR0BfdIXbdrO8aAdN6ANoCEdAkp6qWw/xD3V9lChoBkdAXk9BlcyFf2gHTegDaAhHQJKk0UKzAvd1fZQoaAZHQDLpgssg+yJoB00xAWgIR0CSpqsu3+dcdX2UKGgGR0BjssejmCAdaAdN6ANoCEdAkqxqnR9gGHV9lChoBkdAQEQ9RrJr+GgHTVIBaAhHQJKvDbXYlIF1fZQoaAZHQGN7E1/DtPZoB03oA2gIR0CSsm0NSZSfdX2UKGgGR0BgTypzcRDkaAdN6ANoCEdAkrSOs1baAXV9lChoBkdAaLHYraufVmgHTegDaAhHQJK2xnyup0h1fZQoaAZHQGKFJyyUs4FoB03oA2gIR0CSt2nyup0fdX2UKGgGR0BIUJON5t3waAdL7mgIR0CSusHmzSkTdX2UKGgGR0BmWeU8mrsCaAdN6ANoCEdAktDfkaMrE3V9lChoBkdAYaxNPgvUSmgHTegDaAhHQJLSEagmJFd1fZQoaAZHQF0pNSIgvDhoB03oA2gIR0CS0lzlLeyidX2UKGgGR0BDdW/8EV32aAdNcQFoCEdAktuce0XxfHV9lChoBkdAY2POmixmkGgHTegDaAhHQJLb7qxC6Yp1fZQoaAZHQFh3FGG21D1oB03oA2gIR0CS3BzTWoWIdX2UKGgGR0BjpNXDFZPmaAdN6ANoCEdAkuFGXw9aEHV9lChoBkdAMnic5Ke05WgHTR8BaAhHQJLhprYXfqJ1fZQoaAZHwCFVXFLnLaFoB0uraAhHQJLmHncL0Bh1fZQoaAZHQF7bOzY287JoB03oA2gIR0CS7E0ihWYGdX2UKGgGR8Ao5khA4XGfaAdNNQFoCEdAku7tdAxBV3V9lChoBkdAYbEwpvxYrGgHTegDaAhHQJLvPi0fHPx1fZQoaAZHQGC7i/XXiBJoB03oA2gIR0CS95t+CsfadX2UKGgGR0Bb9BgE2YOUaAdN6ANoCEdAkvloetCAtnV9lChoBkdAO3h+SbH6uWgHTTQBaAhHQJL8Fk078vV1fZQoaAZHQGfOpkwvg3toB03oA2gIR0CTAavl2eQNdX2UKGgGR0BC9pFb3XZoaAdLqWgIR0CTA5TUy57PdX2UKGgGR7/72wiaAnUlaAdNMgFoCEdAkwQ4KtxMnXV9lChoBkdAYu+tTUAks2gHTegDaAhHQJMG8YHgP3B1fZQoaAZHQFtM8hLXcxloB03oA2gIR0CTCT6UaAFxdX2UKGgGR0BkjQphF3INaAdN6ANoCEdAkwn9Jrcj7nV9lChoBkdAXW95Sm65G2gHTegDaAhHQJMNiFFlTWJ1fZQoaAZHQGDFdVvMr3FoB03oA2gIR0CTI8GJN0vHdX2UKGgGR0Bgi4iNbTttaAdN6ANoCEdAkyQooiLVF3V9lChoBkdAOu7aZhKDkGgHS/1oCEdAkyWfJmuklHV9lChoBkdAYHkq6OHWSWgHTegDaAhHQJMuzdP+GXZ1fZQoaAZHQFq5O7QLNOdoB03oA2gIR0CTNRYv38GcdX2UKGgGR0BkV/cYZVGTaAdN6ANoCEdAkzWGki2UjnV9lChoBkdAY/gzhxYJV2gHTegDaAhHQJNBXmr8zhx1fZQoaAZHQGVO7g0j1PFoB03oA2gIR0CTRHkC3gDSdX2UKGgGR0BnrMrGza9LaAdN6ANoCEdAk0wJYs/Y8XV9lChoBkdAW9QqOLiuMmgHTegDaAhHQJNShv60pmV1fZQoaAZHQDtea+evpyJoB0vzaAhHQJNYozZYgaF1fZQoaAZHQGNMhAGB4D9oB03oA2gIR0CTWnG8274BdX2UKGgGR0Bhr2mm+CbuaAdN6ANoCEdAk12Vv/BFeHV9lChoBkdAZsYEq2Bre2gHTegDaAhHQJNgOjWTX8R1fZQoaAZHQGIOuyE+PiloB03oA2gIR0CTYnvOyE+QdX2UKGgGR0Bgj3WJ79hraAdN6ANoCEdAk2MbbxmTT3V9lChoBkdAZWlIPK+zt2gHTegDaAhHQJNmJ8lXzUZ1fZQoaAZHQE1Br8BMi8poB0vyaAhHQJNoo1CPZIx1fZQoaAZHQGY0Y7ihnJ1oB03oA2gIR0CTaZExZdOZdX2UKGgGR0BlDCQ9zOopaAdN6ANoCEdAk2nZr56+nXV9lChoBkfAJD3zDn/1hGgHS49oCEdAk2n+OjqOcXV9lChoBkdAYTAKcd5prWgHTegDaAhHQJN8WGucME11fZQoaAZHQGHPDMV1wHZoB03oA2gIR0CThUZezD4ydX2UKGgGR0BgVT70nPVvaAdN6ANoCEdAk4sugte2NXV9lChoBkdAYk4qiGnGbWgHTegDaAhHQJOLk0GeMAF1fZQoaAZHQEarxDLKV6hoB007AWgIR0CTjuQhOgxrdX2UKGgGR0Bk8jy8SPELaAdN6ANoCEdAk5W3WjGkvnV9lChoBkdAZTv5eqrBCWgHTegDaAhHQJOe9sZYPoV1fZQoaAZHQGQDoYWLxZxoB03oA2gIR0CTpAcghbGFdX2UKGgGR0BoU4D1XeWOaAdN6ANoCEdAk6hpUYKpk3V9lChoBkdAYeTTP0I1L2gHTegDaAhHQJOsxyS3b211fZQoaAZHQFvJg0TDfm9oB03oA2gIR0CTr4gogFHKdX2UKGgGR0BgRUUZeiSJaAdN6ANoCEdAk7OC4Bmwq3V9lChoBkdAQCfNPgvUSmgHTT8BaAhHQJO0ykUKzAx1fZQoaAZHQGOzuejEehhoB03oA2gIR0CTuBRzijtYdX2UKGgGR0BlGuARTS9eaAdN6ANoCEdAk7qfxQSBb3V9lChoBkdAXs8pEx7AtWgHTegDaAhHQJO7izPa+N91fZQoaAZHQGA8UrCm/FloB03oA2gIR0CTu81LrX18dX2UKGgGR0BdQdrO7g89aAdN6ANoCEdAk7zEgKWszXVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVdgIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoCIwCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoESiWCAAAAAAAAAABAQEBAQEBAZRoFUsIhZRoGXSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBEoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaAtLCIWUaBl0lFKUjARoaWdolGgRKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgLSwiFlGgZdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV2wAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCmMBWR0eXBllGgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.2.1+cu121", "GPU Enabled": "True", "Numpy": "1.25.2", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x78fd9012c550>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x78fd9012c5e0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x78fd9012c670>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x78fd9012c700>", "_build": "<function ActorCriticPolicy._build at 0x78fd9012c790>", "forward": "<function ActorCriticPolicy.forward at 0x78fd9012c820>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x78fd9012c8b0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x78fd9012c940>", "_predict": "<function ActorCriticPolicy._predict at 0x78fd9012c9d0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x78fd9012ca60>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x78fd9012caf0>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x78fd9012cb80>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x78fd900ce440>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1710866101321275063, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAJoE/zwp+A+6j7XDutFvB7WlDqe6UBPnOQAAgD8AAIA/MxnXvI+2QbqgPF+6If5wtgdvC7qKWn85AACAPwAAgD+a6W+8XPtwuuNsKLoREja1uqFWO9b/RDkAAIA/AACAP2aMdjzDoT+68/tgO3OfM7aabsi65tQstQAAgD8AAIA/zVDgPLheorkiWCW7WujJtd8lFbs/CkY6AACAPwAAgD8AEVg9wxkQupsh1zswJiu2yKoXuqTvK7UAAIA/AACAP7OpGj1cFz26+f+jukRiLrQDKye70qe/OQAAgD8AAIA/mu6ePFxrN7qGcPk5HdZJtrB5Zrp4bw65AACAPwAAgD9mJI09FKyFuoXlTLmTpkC0f8+cORLPbjgAAIA/AACAP02CYT1cX1u6033TuiZCgbWDdQW7Fo31OQAAgD8AAIA/5resPfaEb7pDw9e5iY2mtgJl/bkeEv44AACAPwAAgD8AEL47SDeGurpDSLozr2g1aVGUujOBZTkAAIA/AACAP7MDGj1h3oI+I7HivDbpk74JNVu9Ft8tPAAAAAAAAAAAZoCMPY8mAroB85O7xoXqtpZ7i7swZ7E6AACAPwAAgD/Nvx49rvmJuhWzJbk87g+08HGxuRyvQDgAAIA/AACAPwBWRrwD+ng9JVKMPQakQL4hcOc9/jlBvQAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQGW6eF10T12MAWyUTegDjAF0lEdAlgPjv3JxN3V9lChoBkdAYmtwH7gsLGgHTegDaAhHQJYHdpztCzF1fZQoaAZHQEA59d/rjYJoB00ZAWgIR0CWHt31jAi3dX2UKGgGR0BhKSG8EmpmaAdN6ANoCEdAlh8sVtXPq3V9lChoBkdAOik3sHB1tGgHTToBaAhHQJYr0qVhTfl1fZQoaAZHQGSfldcB2fVoB03oA2gIR0CWMIMxGlQ/dX2UKGgGR0BjEKXdCVrzaAdN6ANoCEdAljKlLOAy23V9lChoBkdAZKDLnLaEjGgHTegDaAhHQJYy/pgTh5x1fZQoaAZHQGKgenqFAVxoB03oA2gIR0CWNY4VRDTjdX2UKGgGR0BlAb0QK8cuaAdN6ANoCEdAljhB6fJ3gXV9lChoBkdAYjKHmA9V3mgHTegDaAhHQJY5kLRa5gB1fZQoaAZHQGVe7oKUmlZoB03oA2gIR0CWPUTR6WxAdX2UKGgGR0Bev7RSgoPTaAdN6ANoCEdAlj47dWQwK3V9lChoBkdAYDZcnE2pAGgHTegDaAhHQJY/Q+4b0e51fZQoaAZHQF3Bgr6LwWpoB03oA2gIR0CWR2bwjMV2dX2UKGgGR0BicAC+10DEaAdN6ANoCEdAlkkap1ie/nV9lChoBkdAYVFiIcinpGgHTegDaAhHQJZQHrIHTql1fZQoaAZHQGOPMbNr0rdoB03oA2gIR0CWUzQRwqAjdX2UKGgGR0BhGetITXaraAdN6ANoCEdAlm/kgOjIrHV9lChoBkdAYaz6cAimmGgHTegDaAhHQJZwOJqIrOJ1fZQoaAZHQGO1lkhA4XJoB03oA2gIR0CWfRARkEs8dX2UKGgGR0BlJ/SncclxaAdN6ANoCEdAloErSE12q3V9lChoBkdAY800elsP8WgHTegDaAhHQJaCrP+n62x1fZQoaAZHQGXl84HX2/VoB03oA2gIR0CWgu3dbgTAdX2UKGgGR0A4XLhrFfiQaAdNNgFoCEdAloOAlnh86XV9lChoBkdAWn6oaUA1emgHTegDaAhHQJaEuOyVv/B1fZQoaAZHQGHoHUtqYZ5oB03oA2gIR0CWhuI5HVgAdX2UKGgGR0BmlRew9q1xaAdN6ANoCEdAlogUFB6a9nV9lChoBkdAXPJQ66reZWgHTegDaAhHQJaLcUCaJAN1fZQoaAZHQGPSauW8h9toB03oA2gIR0CWjFZmZmZmdX2UKGgGR0BjEGFUQ04zaAdN6ANoCEdAlo1/ddmg8XV9lChoBkdAZF8XGff4y2gHTegDaAhHQJaXZG5MDfZ1fZQoaAZHQGGTpT/ACXBoB03oA2gIR0CWmS+FDfFadX2UKGgGR0Beo/Uz9CNTaAdN6ANoCEdAlqCuNPxhD3V9lChoBkdAYwve4TbnHWgHTegDaAhHQJajzB1s+FF1fZQoaAZHQGRt4XfqHGloB03oA2gIR0CWwBrJ8v25dX2UKGgGR0Bjf4Xwb2lEaAdN6ANoCEdAls3BWgezU3V9lChoBkdAYyioE0SAY2gHTegDaAhHQJbRphoduHh1fZQoaAZHQGM7ktdzGPxoB03oA2gIR0CW0xFr2xptdX2UKGgGR0BmIA+OfdylaAdN6ANoCEdAltNRH5Jsf3V9lChoBkdAZOXWzWwu/WgHTegDaAhHQJbT3OVxCIF1fZQoaAZHQGOany/bj95oB03oA2gIR0CW1PCbc45tdX2UKGgGR0Bh5jM3ZPEbaAdN6ANoCEdAltb98Rcu8XV9lChoBkdAY0qVh1DBuWgHTegDaAhHQJbYB7F85S51fZQoaAZHQGIcbSRbKRxoB03oA2gIR0CW2vY8+zMSdX2UKGgGR0BllRBE8aGYaAdN6ANoCEdAltuzmfXf7HV9lChoBkdAYPjoQnQY12gHTegDaAhHQJbcjkvK2a51fZQoaAZHQGTY8580DU5oB03oA2gIR0CW48Sq2jO+dX2UKGgGR0BkDTposZpBaAdN6ANoCEdAluVyGetjkXV9lChoBkdAYOt9Wp6yB2gHTegDaAhHQJbtcWTHKfZ1fZQoaAZHQF290SRKYiRoB03oA2gIR0CW8Z8cdYGMdX2UKGgGR0Bmn5IxxkupaAdN6ANoCEdAlw43yZrpJXV9lChoBkdAZa42itaIN2gHTegDaAhHQJcdoFKTSst1fZQoaAZHQF3ischkiEBoB03oA2gIR0CXI18KohpydX2UKGgGR0Bj355cC5mRaAdN6ANoCEdAlyUAiFCb+nV9lChoBkdAZFmUC7sfJWgHTegDaAhHQJclRGDtgKF1fZQoaAZHQGZDVNpM6BBoB03oA2gIR0CXJduW8h9tdX2UKGgGR0BHGq+rU9ZBaAdNNAFoCEdAlyYN9H+ZPXV9lChoBkdAYhnW5H3DemgHTegDaAhHQJcnHhrFfiR1fZQoaAZHQGYhUGu9vjxoB03oA2gIR0CXKUBX0XgtdX2UKGgGR0BhPGj7ALy+aAdN6ANoCEdAlypaMrEtNHV9lChoBkdAZhezN2TxG2gHTegDaAhHQJctgna37UJ1fZQoaAZHQF8AtmL9/BpoB03oA2gIR0CXLlhrWRRudX2UKGgGR0Bbi3meUY8/aAdN6ANoCEdAly9MQiA2AHV9lChoBkdAW5gir1dxAGgHTegDaAhHQJc3wAPuogp1fZQoaAZHQGcV67EpAlhoB03oA2gIR0CXOZ2eQMhHdX2UKGgGR0BjRV9Sde6aaAdN6ANoCEdAl0Egiu+yq3V9lChoBkdAYYpXbuc+aGgHTegDaAhHQJdEkkNWluZ1fZQoaAZHQGKKv557gKpoB03oA2gIR0CXcdIInjQzdX2UKGgGR0BkAw++ueSTaAdN6ANoCEdAl3ZRas6q83V9lChoBkdAZwjFId2gWmgHTegDaAhHQJd30FmnO0N1fZQoaAZHQF97KDkELYxoB03oA2gIR0CXeBCPIXCTdX2UKGgGR0BnbT9KmKqGaAdN6ANoCEdAl3ibmp2lmHV9lChoBkdAYUDrP+n622gHTegDaAhHQJd4yDwpe/p1fZQoaAZHQGOvgtFrl/9oB03oA2gIR0CXedUc4o7WdX2UKGgGR0BgynVbzK9xaAdN6ANoCEdAl3ytTkyULXV9lChoBkdAZOCFkhA4XGgHTegDaAhHQJd+EOEug6F1fZQoaAZHQGW8O3+dbxFoB03oA2gIR0CXgeN+so2GdX2UKGgGR0BfL4acZtN0aAdN6ANoCEdAl4K6IznA7HV9lChoBkdAZcsI1tO2zGgHTegDaAhHQJeDnLeQ+2V1fZQoaAZHQGFo5jYqXnhoB03oA2gIR0CXiyWdVea8dX2UKGgGR0BnskdPtUn5aAdN6ANoCEdAl4zNrO7g9HV9lChoBkdAYMbo0Q9RrWgHTegDaAhHQJeT2Oq//Nt1fZQoaAZHQGSfi8WbgCRoB03oA2gIR0CXlvJ0nw5OdX2UKGgGR0Bl7eom5UcXaAdN6ANoCEdAl8TP0qYqonV9lChoBkdAXVe0Xxe9jGgHTegDaAhHQJfKAetCAtp1fZQoaAZHQGSLrS3LFGZoB03oA2gIR0CXy93i704BdX2UKGgGR0BlqQntv4ucaAdN6ANoCEdAl8wqaPS2IHV9lChoBkdAZ7+rZrYXf2gHTegDaAhHQJfM23ocJdB1fZQoaAZHQGQk3juKGcpoB03oA2gIR0CXzQ8D0UXYdX2UKGgGR0BhcJO8CgbqaAdN6ANoCEdAl84kGqxTsXV9lChoBkdAYqDs0HhS+GgHTegDaAhHQJfQaBas6q91fZQoaAZHQGMhfzJ6po9oB03oA2gIR0CX0Zt5D7ZWdX2UKGgGR0BgTJGhEjPfaAdN6ANoCEdAl9UcJMQEp3V9lChoBkdAZBHMr3CbdGgHTegDaAhHQJfWAZ88cMp1fZQoaAZHQGYsk8A7xNJoB03oA2gIR0CX1wPiDM/ydX2UKGgGR0BgWHe1rqMWaAdN6ANoCEdAl+FtadMCcXV9lChoBkdAYoCMglnh9GgHTegDaAhHQJfjJ2IO6NF1fZQoaAZHQGYwE0zj3mFoB03oA2gIR0CX6kJyQxN7dX2UKGgGR0BgizxXnyNGaAdN6ANoCEdAl+1Oqebut3VlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVdgIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoCIwCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoESiWCAAAAAAAAAABAQEBAQEBAZRoFUsIhZRoGXSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBEoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaAtLCIWUaBl0lFKUjARoaWdolGgRKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgLSwiFlGgZdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV2wAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCmMBWR0eXBllGgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.2.1+cu121", "GPU Enabled": "True", "Numpy": "1.25.2", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
ppo-LunarLander-v2.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74f7a10b194904c82940ff0d0fd973ccbc283abad55955a3d17d669c3d5f6ac4
|
3 |
+
size 148088
|
ppo-LunarLander-v2/data
CHANGED
@@ -4,20 +4,20 @@
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
11 |
-
"_build": "<function ActorCriticPolicy._build at
|
12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
-
"_abc_impl": "<_abc._abc_data object at
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {},
|
@@ -26,12 +26,12 @@
|
|
26 |
"_num_timesteps_at_start": 0,
|
27 |
"seed": null,
|
28 |
"action_noise": null,
|
29 |
-
"start_time":
|
30 |
"learning_rate": 0.0003,
|
31 |
"tensorboard_log": null,
|
32 |
"_last_obs": {
|
33 |
":type:": "<class 'numpy.ndarray'>",
|
34 |
-
":serialized:": "
|
35 |
},
|
36 |
"_last_episode_starts": {
|
37 |
":type:": "<class 'numpy.ndarray'>",
|
@@ -45,7 +45,7 @@
|
|
45 |
"_stats_window_size": 100,
|
46 |
"ep_info_buffer": {
|
47 |
":type:": "<class 'collections.deque'>",
|
48 |
-
":serialized:": "
|
49 |
},
|
50 |
"ep_success_buffer": {
|
51 |
":type:": "<class 'collections.deque'>",
|
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x78fd9012c550>",
|
8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x78fd9012c5e0>",
|
9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x78fd9012c670>",
|
10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x78fd9012c700>",
|
11 |
+
"_build": "<function ActorCriticPolicy._build at 0x78fd9012c790>",
|
12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x78fd9012c820>",
|
13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x78fd9012c8b0>",
|
14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x78fd9012c940>",
|
15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x78fd9012c9d0>",
|
16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x78fd9012ca60>",
|
17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x78fd9012caf0>",
|
18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x78fd9012cb80>",
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
+
"_abc_impl": "<_abc._abc_data object at 0x78fd900ce440>"
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {},
|
|
|
26 |
"_num_timesteps_at_start": 0,
|
27 |
"seed": null,
|
28 |
"action_noise": null,
|
29 |
+
"start_time": 1710866101321275063,
|
30 |
"learning_rate": 0.0003,
|
31 |
"tensorboard_log": null,
|
32 |
"_last_obs": {
|
33 |
":type:": "<class 'numpy.ndarray'>",
|
34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAJoE/zwp+A+6j7XDutFvB7WlDqe6UBPnOQAAgD8AAIA/MxnXvI+2QbqgPF+6If5wtgdvC7qKWn85AACAPwAAgD+a6W+8XPtwuuNsKLoREja1uqFWO9b/RDkAAIA/AACAP2aMdjzDoT+68/tgO3OfM7aabsi65tQstQAAgD8AAIA/zVDgPLheorkiWCW7WujJtd8lFbs/CkY6AACAPwAAgD8AEVg9wxkQupsh1zswJiu2yKoXuqTvK7UAAIA/AACAP7OpGj1cFz26+f+jukRiLrQDKye70qe/OQAAgD8AAIA/mu6ePFxrN7qGcPk5HdZJtrB5Zrp4bw65AACAPwAAgD9mJI09FKyFuoXlTLmTpkC0f8+cORLPbjgAAIA/AACAP02CYT1cX1u6033TuiZCgbWDdQW7Fo31OQAAgD8AAIA/5resPfaEb7pDw9e5iY2mtgJl/bkeEv44AACAPwAAgD8AEL47SDeGurpDSLozr2g1aVGUujOBZTkAAIA/AACAP7MDGj1h3oI+I7HivDbpk74JNVu9Ft8tPAAAAAAAAAAAZoCMPY8mAroB85O7xoXqtpZ7i7swZ7E6AACAPwAAgD/Nvx49rvmJuhWzJbk87g+08HGxuRyvQDgAAIA/AACAPwBWRrwD+ng9JVKMPQakQL4hcOc9/jlBvQAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
35 |
},
|
36 |
"_last_episode_starts": {
|
37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
45 |
"_stats_window_size": 100,
|
46 |
"ep_info_buffer": {
|
47 |
":type:": "<class 'collections.deque'>",
|
48 |
+
":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQGW6eF10T12MAWyUTegDjAF0lEdAlgPjv3JxN3V9lChoBkdAYmtwH7gsLGgHTegDaAhHQJYHdpztCzF1fZQoaAZHQEA59d/rjYJoB00ZAWgIR0CWHt31jAi3dX2UKGgGR0BhKSG8EmpmaAdN6ANoCEdAlh8sVtXPq3V9lChoBkdAOik3sHB1tGgHTToBaAhHQJYr0qVhTfl1fZQoaAZHQGSfldcB2fVoB03oA2gIR0CWMIMxGlQ/dX2UKGgGR0BjEKXdCVrzaAdN6ANoCEdAljKlLOAy23V9lChoBkdAZKDLnLaEjGgHTegDaAhHQJYy/pgTh5x1fZQoaAZHQGKgenqFAVxoB03oA2gIR0CWNY4VRDTjdX2UKGgGR0BlAb0QK8cuaAdN6ANoCEdAljhB6fJ3gXV9lChoBkdAYjKHmA9V3mgHTegDaAhHQJY5kLRa5gB1fZQoaAZHQGVe7oKUmlZoB03oA2gIR0CWPUTR6WxAdX2UKGgGR0Bev7RSgoPTaAdN6ANoCEdAlj47dWQwK3V9lChoBkdAYDZcnE2pAGgHTegDaAhHQJY/Q+4b0e51fZQoaAZHQF3Bgr6LwWpoB03oA2gIR0CWR2bwjMV2dX2UKGgGR0BicAC+10DEaAdN6ANoCEdAlkkap1ie/nV9lChoBkdAYVFiIcinpGgHTegDaAhHQJZQHrIHTql1fZQoaAZHQGOPMbNr0rdoB03oA2gIR0CWUzQRwqAjdX2UKGgGR0BhGetITXaraAdN6ANoCEdAlm/kgOjIrHV9lChoBkdAYaz6cAimmGgHTegDaAhHQJZwOJqIrOJ1fZQoaAZHQGO1lkhA4XJoB03oA2gIR0CWfRARkEs8dX2UKGgGR0BlJ/SncclxaAdN6ANoCEdAloErSE12q3V9lChoBkdAY800elsP8WgHTegDaAhHQJaCrP+n62x1fZQoaAZHQGXl84HX2/VoB03oA2gIR0CWgu3dbgTAdX2UKGgGR0A4XLhrFfiQaAdNNgFoCEdAloOAlnh86XV9lChoBkdAWn6oaUA1emgHTegDaAhHQJaEuOyVv/B1fZQoaAZHQGHoHUtqYZ5oB03oA2gIR0CWhuI5HVgAdX2UKGgGR0BmlRew9q1xaAdN6ANoCEdAlogUFB6a9nV9lChoBkdAXPJQ66reZWgHTegDaAhHQJaLcUCaJAN1fZQoaAZHQGPSauW8h9toB03oA2gIR0CWjFZmZmZmdX2UKGgGR0BjEGFUQ04zaAdN6ANoCEdAlo1/ddmg8XV9lChoBkdAZF8XGff4y2gHTegDaAhHQJaXZG5MDfZ1fZQoaAZHQGGTpT/ACXBoB03oA2gIR0CWmS+FDfFadX2UKGgGR0Beo/Uz9CNTaAdN6ANoCEdAlqCuNPxhD3V9lChoBkdAYwve4TbnHWgHTegDaAhHQJajzB1s+FF1fZQoaAZHQGRt4XfqHGloB03oA2gIR0CWwBrJ8v25dX2UKGgGR0Bjf4Xwb2lEaAdN6ANoCEdAls3BWgezU3V9lChoBkdAYyioE0SAY2gHTegDaAhHQJbRphoduHh1fZQoaAZHQGM7ktdzGPxoB03oA2gIR0CW0xFr2xptdX2UKGgGR0BmIA+OfdylaAdN6ANoCEdAltNRH5Jsf3V9lChoBkdAZOXWzWwu/WgHTegDaAhHQJbT3OVxCIF1fZQoaAZHQGOany/bj95oB03oA2gIR0CW1PCbc45tdX2UKGgGR0Bh5jM3ZPEbaAdN6ANoCEdAltb98Rcu8XV9lChoBkdAY0qVh1DBuWgHTegDaAhHQJbYB7F85S51fZQoaAZHQGIcbSRbKRxoB03oA2gIR0CW2vY8+zMSdX2UKGgGR0BllRBE8aGYaAdN6ANoCEdAltuzmfXf7HV9lChoBkdAYPjoQnQY12gHTegDaAhHQJbcjkvK2a51fZQoaAZHQGTY8580DU5oB03oA2gIR0CW48Sq2jO+dX2UKGgGR0BkDTposZpBaAdN6ANoCEdAluVyGetjkXV9lChoBkdAYOt9Wp6yB2gHTegDaAhHQJbtcWTHKfZ1fZQoaAZHQF290SRKYiRoB03oA2gIR0CW8Z8cdYGMdX2UKGgGR0Bmn5IxxkupaAdN6ANoCEdAlw43yZrpJXV9lChoBkdAZa42itaIN2gHTegDaAhHQJcdoFKTSst1fZQoaAZHQF3ischkiEBoB03oA2gIR0CXI18KohpydX2UKGgGR0Bj355cC5mRaAdN6ANoCEdAlyUAiFCb+nV9lChoBkdAZFmUC7sfJWgHTegDaAhHQJclRGDtgKF1fZQoaAZHQGZDVNpM6BBoB03oA2gIR0CXJduW8h9tdX2UKGgGR0BHGq+rU9ZBaAdNNAFoCEdAlyYN9H+ZPXV9lChoBkdAYhnW5H3DemgHTegDaAhHQJcnHhrFfiR1fZQoaAZHQGYhUGu9vjxoB03oA2gIR0CXKUBX0XgtdX2UKGgGR0BhPGj7ALy+aAdN6ANoCEdAlypaMrEtNHV9lChoBkdAZhezN2TxG2gHTegDaAhHQJctgna37UJ1fZQoaAZHQF8AtmL9/BpoB03oA2gIR0CXLlhrWRRudX2UKGgGR0Bbi3meUY8/aAdN6ANoCEdAly9MQiA2AHV9lChoBkdAW5gir1dxAGgHTegDaAhHQJc3wAPuogp1fZQoaAZHQGcV67EpAlhoB03oA2gIR0CXOZ2eQMhHdX2UKGgGR0BjRV9Sde6aaAdN6ANoCEdAl0Egiu+yq3V9lChoBkdAYYpXbuc+aGgHTegDaAhHQJdEkkNWluZ1fZQoaAZHQGKKv557gKpoB03oA2gIR0CXcdIInjQzdX2UKGgGR0BkAw++ueSTaAdN6ANoCEdAl3ZRas6q83V9lChoBkdAZwjFId2gWmgHTegDaAhHQJd30FmnO0N1fZQoaAZHQF97KDkELYxoB03oA2gIR0CXeBCPIXCTdX2UKGgGR0BnbT9KmKqGaAdN6ANoCEdAl3ibmp2lmHV9lChoBkdAYUDrP+n622gHTegDaAhHQJd4yDwpe/p1fZQoaAZHQGOvgtFrl/9oB03oA2gIR0CXedUc4o7WdX2UKGgGR0BgynVbzK9xaAdN6ANoCEdAl3ytTkyULXV9lChoBkdAZOCFkhA4XGgHTegDaAhHQJd+EOEug6F1fZQoaAZHQGW8O3+dbxFoB03oA2gIR0CXgeN+so2GdX2UKGgGR0BfL4acZtN0aAdN6ANoCEdAl4K6IznA7HV9lChoBkdAZcsI1tO2zGgHTegDaAhHQJeDnLeQ+2V1fZQoaAZHQGFo5jYqXnhoB03oA2gIR0CXiyWdVea8dX2UKGgGR0BnskdPtUn5aAdN6ANoCEdAl4zNrO7g9HV9lChoBkdAYMbo0Q9RrWgHTegDaAhHQJeT2Oq//Nt1fZQoaAZHQGSfi8WbgCRoB03oA2gIR0CXlvJ0nw5OdX2UKGgGR0Bl7eom5UcXaAdN6ANoCEdAl8TP0qYqonV9lChoBkdAXVe0Xxe9jGgHTegDaAhHQJfKAetCAtp1fZQoaAZHQGSLrS3LFGZoB03oA2gIR0CXy93i704BdX2UKGgGR0BlqQntv4ucaAdN6ANoCEdAl8wqaPS2IHV9lChoBkdAZ7+rZrYXf2gHTegDaAhHQJfM23ocJdB1fZQoaAZHQGQk3juKGcpoB03oA2gIR0CXzQ8D0UXYdX2UKGgGR0BhcJO8CgbqaAdN6ANoCEdAl84kGqxTsXV9lChoBkdAYqDs0HhS+GgHTegDaAhHQJfQaBas6q91fZQoaAZHQGMhfzJ6po9oB03oA2gIR0CX0Zt5D7ZWdX2UKGgGR0BgTJGhEjPfaAdN6ANoCEdAl9UcJMQEp3V9lChoBkdAZBHMr3CbdGgHTegDaAhHQJfWAZ88cMp1fZQoaAZHQGYsk8A7xNJoB03oA2gIR0CX1wPiDM/ydX2UKGgGR0BgWHe1rqMWaAdN6ANoCEdAl+FtadMCcXV9lChoBkdAYoCMglnh9GgHTegDaAhHQJfjJ2IO6NF1fZQoaAZHQGYwE0zj3mFoB03oA2gIR0CX6kJyQxN7dX2UKGgGR0BgizxXnyNGaAdN6ANoCEdAl+1Oqebut3VlLg=="
|
49 |
},
|
50 |
"ep_success_buffer": {
|
51 |
":type:": "<class 'collections.deque'>",
|
ppo-LunarLander-v2/policy.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 88362
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91fdea2fcac44e6d37403023b71ba8247eddf20b34bfd8a8e1191913963995e9
|
3 |
size 88362
|
ppo-LunarLander-v2/policy.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 43762
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d46b16df847286662215a4a9dbfd9a378055ca67da15a8b90f6f25a15acc6010
|
3 |
size 43762
|
replay.mp4
CHANGED
Binary files a/replay.mp4 and b/replay.mp4 differ
|
|
results.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"mean_reward":
|
|
|
1 |
+
{"mean_reward": 258.5701146, "std_reward": 20.65045160115703, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-03-19T16:56:45.706726"}
|