Theaveas commited on
Commit
c41aa32
1 Parent(s): c35c35d

Upload PPO Mountain trained model

Browse files
.gitattributes CHANGED
@@ -25,3 +25,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ *.mp4 filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: stable-baselines3
3
+ tags:
4
+ - MountainCar-v0
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - stable-baselines3
8
+ model-index:
9
+ - name: PPO
10
+ results:
11
+ - metrics:
12
+ - type: mean_reward
13
+ value: -97.20 +/- 7.12
14
+ name: mean_reward
15
+ task:
16
+ type: reinforcement-learning
17
+ name: reinforcement-learning
18
+ dataset:
19
+ name: MountainCar-v0
20
+ type: MountainCar-v0
21
+ ---
22
+
23
+ # **PPO** Agent playing **MountainCar-v0**
24
+ This is a trained model of a **PPO** agent playing **MountainCar-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
25
+
26
+ ## Usage (with Stable-baselines3)
27
+ TODO: Add your code
28
+
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x138119c10>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x138119ca0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x138119d30>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x138119dc0>", "_build": "<function ActorCriticPolicy._build at 0x138119e50>", "forward": "<function ActorCriticPolicy.forward at 0x138119ee0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x138119f70>", "_predict": "<function ActorCriticPolicy._predict at 0x138120040>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x1381200d0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x138120160>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x1381201f0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x12fec56c0>"}, "verbose": 1, "policy_kwargs": {}, "observation_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVYwEAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLAoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAACamZm/KVyPvZRoCksChZSMAUOUdJRSlIwEaGlnaJRoEiiWCAAAAAAAAACamRk/KVyPPZRoCksChZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolgIAAAAAAAAAAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLAoWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYCAAAAAAAAAAEBlGghSwKFlGgVdJRSlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "_shape": [2], "low": "[-1.2 -0.07]", "high": "[0.6 0.07]", "bounded_below": "[ True True]", "bounded_above": "[ True True]", "_np_random": null}, "action_space": {":type:": "<class 'gym.spaces.discrete.Discrete'>", ":serialized:": "gAWVggAAAAAAAACME2d5bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpRLA4wGX3NoYXBllCmMBWR0eXBllIwFbnVtcHmUaAeTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYowKX25wX3JhbmRvbZROdWIu", "n": 3, "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 1, "num_timesteps": 1000000, "_total_timesteps": 3000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1651836900.860854, "learning_rate": 0.0003, "tensorboard_log": "logs", "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV/wIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwNX2J1aWx0aW5fdHlwZZSTlIwKTGFtYmRhVHlwZZSFlFKUKGgCjAhDb2RlVHlwZZSFlFKUKEsBSwBLAEsBSwFLE0MEiABTAJROhZQpjAFflIWUjGcvVXNlcnMvdGhlYXZlYXNzby9taW5pZm9yZ2UzL2VudnMvaW50cm9STC9saWIvcHl0aG9uMy45L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxnL1VzZXJzL3RoZWF2ZWFzc28vbWluaWZvcmdlMy9lbnZzL2ludHJvUkwvbGliL3B5dGhvbjMuOS9zaXRlLXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGggfZR9lChoF2gOjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoGIwHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/M6kqMFUyYYWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="}, "_last_obs": null, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdAAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYBAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwGFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": 0.6668586666666667, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHwFoAAAAAAACMAWyUS2iMAXSUR0BzNqJdjXnRdX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzNtSEUTL4dX2UKGgGR8BWwAAAAAAAaAdLW2gIR0BzNwIF/x2CdX2UKGgGR8BWwAAAAAAAaAdLW2gIR0BzNy/O+qR2dX2UKGgGR8BYgAAAAAAAaAdLYmgIR0BzN2CHymQ9dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzN5SBK+SKdX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzN8aqCHymdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzN/D+BH09dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzOCMqBmPHdX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzOFxWDHwPdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzOJPXTVlPdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzOMq3EyckdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzPsA0bcXWdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzPvSH/LkkdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzPyZ0CA+ZdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzP1gXuVopdX2UKGgGR8BbwAAAAAAAaAdLb2gIR0BzP41k1/DtdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzP78yeqaPdX2UKGgGR8BbAAAAAAAAaAdLbGgIR0BzP/MdLg4wdX2UKGgGR8BmoAAAAAAAaAdLtWgIR0BzQErNGEwndX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzQH1pTMq0dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQLEaVD8cdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzQNs9B8hLdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzQQ593KSxdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzQTicXm/4dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQWxeLNwBdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzQZaB7NSqdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzQcq+ajN7dX2UKGgGR8BXgAAAAAAAaAdLXmgIR0BzQfj6vaDgdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQioS+QEIdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzQlIczZYgdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQoNc4YJmdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzSawOe8PGdX2UKGgGR8BngAAAAAAAaAdLvGgIR0BzSgpb2USqdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzSj1HvttzdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzSnRsuWa+dX2UKGgGR8BWAAAAAAAAaAdLWGgIR0BzSqBreqJedX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzStQtSQ5ndX2UKGgGR8BZQAAAAAAAaAdLZWgIR0BzSwVUMoc8dX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzSzm0VrRCdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzS2/tY0VKdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzS6Lm6oVEdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzS9iH6/IsdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzTA4WDYh/dX2UKGgGR8BVwAAAAAAAaAdLV2gIR0BzTDi97F85dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzTG1stTUBdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzTKEmICU5dX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzTMvXbuc+dX2UKGgGR8BVgAAAAAAAaAdLVmgIR0BzTPXjENvwdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzTSvkili0dX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzTVdB0ITodX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzTYsI3R5UdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzVNDSgGr0dX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzVP4SHuZ1dX2UKGgGR8BdAAAAAAAAaAdLdGgIR0BzVTkp7TlUdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzVWQ0XP7fdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzVZsyi22HdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzVdDUmUnpdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzVgXl8w6AdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzVjoTwlSkdX2UKGgGR8BVgAAAAAAAaAdLVmgIR0BzVmXTmW+odX2UKGgGR8BWQAAAAAAAaAdLWWgIR0BzVpMVUModdX2UKGgGR8BXgAAAAAAAaAdLXmgIR0BzVsLncL0BdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzVvhVENONdX2UKGgGR8BVwAAAAAAAaAdLV2gIR0BzVyOo5xR3dX2UKGgGR8BUwAAAAAAAaAdLU2gIR0BzV04ecQRPdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzV4LWqcVhdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzV60pmVZ+dX2UKGgGR8BWQAAAAAAAaAdLWWgIR0BzV9uGbkOqdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzWA99tuUEdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzWEIeHSF5dX2UKGgGR8BXAAAAAAAAaAdLXGgIR0BzWHB9Cu2adX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzWKJ53TuwdX2UKGgGR8BVgAAAAAAAaAdLVmgIR0BzXv1zySV4dX2UKGgGR8BeQAAAAAAAaAdLeWgIR0BzXzoePq9odX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzX21+iJwbdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzX6GGmDUWdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzX9T987ZGdX2UKGgGR8BWgAAAAAAAaAdLWmgIR0BzYAH5aePJdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYDXjENvwdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYGm3vx6OdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYJ1uBMBZdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYNC6Ymb9dX2UKGgGR8BWAAAAAAAAaAdLWGgIR0BzYPxd6cAjdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYTBBRhttdX2UKGgGR8BVwAAAAAAAaAdLV2gIR0BzYVtIkJKKdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzYYUVSGahdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYbjHXEqEdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzYeKUFB6bdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzYhbX6InCdX2UKGgGR8BYQAAAAAAAaAdLYWgIR0BzYkfOlfqpdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYnp/wy6+dX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzYqNzbN8mdX2UKGgGR8BXwAAAAAAAaAdLX2gIR0BzYtFa0QbudX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzaLK/20zCdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzaOqCHymRdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzaR+kP+XJdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzaVVGTcIrdX2UKGgGR8BXAAAAAAAAaAdLXGgIR0BzaYRqXWvsdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzabqB3A2ydWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 9770, "n_steps": 2048, "gamma": 0.99, "gae_lambda": 0.95, "ent_coef": 0.0, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 10, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV/wIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwNX2J1aWx0aW5fdHlwZZSTlIwKTGFtYmRhVHlwZZSFlFKUKGgCjAhDb2RlVHlwZZSFlFKUKEsBSwBLAEsBSwFLE0MEiABTAJROhZQpjAFflIWUjGcvVXNlcnMvdGhlYXZlYXNzby9taW5pZm9yZ2UzL2VudnMvaW50cm9STC9saWIvcHl0aG9uMy45L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxnL1VzZXJzL3RoZWF2ZWFzc28vbWluaWZvcmdlMy9lbnZzL2ludHJvUkwvbGliL3B5dGhvbjMuOS9zaXRlLXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGggfZR9lChoF2gOjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoGIwHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/yZmZmZmZmoWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "system_info": {"OS": "macOS-12.3.1-arm64-arm-64bit Darwin Kernel Version 21.4.0: Fri Mar 18 00:46:32 PDT 2022; root:xnu-8020.101.4~15/RELEASE_ARM64_T6000", "Python": "3.9.12", "Stable-Baselines3": "1.5.0", "PyTorch": "1.11.0", "GPU Enabled": "False", "Numpy": "1.22.3", "Gym": "0.21.0"}}
ppo-mountaincar-v0.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:568ecec89897b61ced38c6a42afbff1b1ae41a174c965ba5eb0efe445b5d8fc0
3
+ size 131199
ppo-mountaincar-v0/_stable_baselines3_version ADDED
@@ -0,0 +1 @@
 
 
1
+ 1.5.0
ppo-mountaincar-v0/data ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "policy_class": {
3
+ ":type:": "<class 'abc.ABCMeta'>",
4
+ ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
5
+ "__module__": "stable_baselines3.common.policies",
6
+ "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
7
+ "__init__": "<function ActorCriticPolicy.__init__ at 0x138119c10>",
8
+ "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x138119ca0>",
9
+ "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x138119d30>",
10
+ "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x138119dc0>",
11
+ "_build": "<function ActorCriticPolicy._build at 0x138119e50>",
12
+ "forward": "<function ActorCriticPolicy.forward at 0x138119ee0>",
13
+ "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x138119f70>",
14
+ "_predict": "<function ActorCriticPolicy._predict at 0x138120040>",
15
+ "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x1381200d0>",
16
+ "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x138120160>",
17
+ "predict_values": "<function ActorCriticPolicy.predict_values at 0x1381201f0>",
18
+ "__abstractmethods__": "frozenset()",
19
+ "_abc_impl": "<_abc._abc_data object at 0x12fec56c0>"
20
+ },
21
+ "verbose": 1,
22
+ "policy_kwargs": {},
23
+ "observation_space": {
24
+ ":type:": "<class 'gym.spaces.box.Box'>",
25
+ ":serialized:": "gAWVYwEAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLAoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAACamZm/KVyPvZRoCksChZSMAUOUdJRSlIwEaGlnaJRoEiiWCAAAAAAAAACamRk/KVyPPZRoCksChZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolgIAAAAAAAAAAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLAoWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYCAAAAAAAAAAEBlGghSwKFlGgVdJRSlIwKX25wX3JhbmRvbZROdWIu",
26
+ "dtype": "float32",
27
+ "_shape": [
28
+ 2
29
+ ],
30
+ "low": "[-1.2 -0.07]",
31
+ "high": "[0.6 0.07]",
32
+ "bounded_below": "[ True True]",
33
+ "bounded_above": "[ True True]",
34
+ "_np_random": null
35
+ },
36
+ "action_space": {
37
+ ":type:": "<class 'gym.spaces.discrete.Discrete'>",
38
+ ":serialized:": "gAWVggAAAAAAAACME2d5bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpRLA4wGX3NoYXBllCmMBWR0eXBllIwFbnVtcHmUaAeTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYowKX25wX3JhbmRvbZROdWIu",
39
+ "n": 3,
40
+ "_shape": [],
41
+ "dtype": "int64",
42
+ "_np_random": null
43
+ },
44
+ "n_envs": 1,
45
+ "num_timesteps": 1000000,
46
+ "_total_timesteps": 3000000,
47
+ "_num_timesteps_at_start": 0,
48
+ "seed": null,
49
+ "action_noise": null,
50
+ "start_time": 1651836900.860854,
51
+ "learning_rate": 0.0003,
52
+ "tensorboard_log": "logs",
53
+ "lr_schedule": {
54
+ ":type:": "<class 'function'>",
55
+ ":serialized:": "gAWV/wIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwNX2J1aWx0aW5fdHlwZZSTlIwKTGFtYmRhVHlwZZSFlFKUKGgCjAhDb2RlVHlwZZSFlFKUKEsBSwBLAEsBSwFLE0MEiABTAJROhZQpjAFflIWUjGcvVXNlcnMvdGhlYXZlYXNzby9taW5pZm9yZ2UzL2VudnMvaW50cm9STC9saWIvcHl0aG9uMy45L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxnL1VzZXJzL3RoZWF2ZWFzc28vbWluaWZvcmdlMy9lbnZzL2ludHJvUkwvbGliL3B5dGhvbjMuOS9zaXRlLXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGggfZR9lChoF2gOjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoGIwHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/M6kqMFUyYYWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
56
+ },
57
+ "_last_obs": null,
58
+ "_last_episode_starts": {
59
+ ":type:": "<class 'numpy.ndarray'>",
60
+ ":serialized:": "gAWVdAAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYBAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwGFlIwBQ5R0lFKULg=="
61
+ },
62
+ "_last_original_obs": null,
63
+ "_episode_num": 0,
64
+ "use_sde": false,
65
+ "sde_sample_freq": -1,
66
+ "_current_progress_remaining": 0.6668586666666667,
67
+ "ep_info_buffer": {
68
+ ":type:": "<class 'collections.deque'>",
69
+ ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHwFoAAAAAAACMAWyUS2iMAXSUR0BzNqJdjXnRdX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzNtSEUTL4dX2UKGgGR8BWwAAAAAAAaAdLW2gIR0BzNwIF/x2CdX2UKGgGR8BWwAAAAAAAaAdLW2gIR0BzNy/O+qR2dX2UKGgGR8BYgAAAAAAAaAdLYmgIR0BzN2CHymQ9dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzN5SBK+SKdX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzN8aqCHymdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzN/D+BH09dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzOCMqBmPHdX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzOFxWDHwPdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzOJPXTVlPdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzOMq3EyckdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzPsA0bcXWdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzPvSH/LkkdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzPyZ0CA+ZdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzP1gXuVopdX2UKGgGR8BbwAAAAAAAaAdLb2gIR0BzP41k1/DtdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzP78yeqaPdX2UKGgGR8BbAAAAAAAAaAdLbGgIR0BzP/MdLg4wdX2UKGgGR8BmoAAAAAAAaAdLtWgIR0BzQErNGEwndX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzQH1pTMq0dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQLEaVD8cdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzQNs9B8hLdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzQQ593KSxdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzQTicXm/4dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQWxeLNwBdX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzQZaB7NSqdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzQcq+ajN7dX2UKGgGR8BXgAAAAAAAaAdLXmgIR0BzQfj6vaDgdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQioS+QEIdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzQlIczZYgdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzQoNc4YJmdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzSawOe8PGdX2UKGgGR8BngAAAAAAAaAdLvGgIR0BzSgpb2USqdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzSj1HvttzdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzSnRsuWa+dX2UKGgGR8BWAAAAAAAAaAdLWGgIR0BzSqBreqJedX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzStQtSQ5ndX2UKGgGR8BZQAAAAAAAaAdLZWgIR0BzSwVUMoc8dX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzSzm0VrRCdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzS2/tY0VKdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzS6Lm6oVEdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzS9iH6/IsdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzTA4WDYh/dX2UKGgGR8BVwAAAAAAAaAdLV2gIR0BzTDi97F85dX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzTG1stTUBdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzTKEmICU5dX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzTMvXbuc+dX2UKGgGR8BVgAAAAAAAaAdLVmgIR0BzTPXjENvwdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzTSvkili0dX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzTVdB0ITodX2UKGgGR8BZgAAAAAAAaAdLZmgIR0BzTYsI3R5UdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzVNDSgGr0dX2UKGgGR8BVQAAAAAAAaAdLVWgIR0BzVP4SHuZ1dX2UKGgGR8BdAAAAAAAAaAdLdGgIR0BzVTkp7TlUdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzVWQ0XP7fdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzVZsyi22HdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzVdDUmUnpdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzVgXl8w6AdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzVjoTwlSkdX2UKGgGR8BVgAAAAAAAaAdLVmgIR0BzVmXTmW+odX2UKGgGR8BWQAAAAAAAaAdLWWgIR0BzVpMVUModdX2UKGgGR8BXgAAAAAAAaAdLXmgIR0BzVsLncL0BdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzVvhVENONdX2UKGgGR8BVwAAAAAAAaAdLV2gIR0BzVyOo5xR3dX2UKGgGR8BUwAAAAAAAaAdLU2gIR0BzV04ecQRPdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzV4LWqcVhdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzV60pmVZ+dX2UKGgGR8BWQAAAAAAAaAdLWWgIR0BzV9uGbkOqdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzWA99tuUEdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzWEIeHSF5dX2UKGgGR8BXAAAAAAAAaAdLXGgIR0BzWHB9Cu2adX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzWKJ53TuwdX2UKGgGR8BVgAAAAAAAaAdLVmgIR0BzXv1zySV4dX2UKGgGR8BeQAAAAAAAaAdLeWgIR0BzXzoePq9odX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzX21+iJwbdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzX6GGmDUWdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzX9T987ZGdX2UKGgGR8BWgAAAAAAAaAdLWmgIR0BzYAH5aePJdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYDXjENvwdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYGm3vx6OdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYJ1uBMBZdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYNC6Ymb9dX2UKGgGR8BWAAAAAAAAaAdLWGgIR0BzYPxd6cAjdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYTBBRhttdX2UKGgGR8BVwAAAAAAAaAdLV2gIR0BzYVtIkJKKdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzYYUVSGahdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYbjHXEqEdX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzYeKUFB6bdX2UKGgGR8BaQAAAAAAAaAdLaWgIR0BzYhbX6InCdX2UKGgGR8BYQAAAAAAAaAdLYWgIR0BzYkfOlfqpdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzYnp/wy6+dX2UKGgGR8BVAAAAAAAAaAdLVGgIR0BzYqNzbN8mdX2UKGgGR8BXwAAAAAAAaAdLX2gIR0BzYtFa0QbudX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzaLK/20zCdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzaOqCHymRdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzaR+kP+XJdX2UKGgGR8BaAAAAAAAAaAdLaGgIR0BzaVVGTcIrdX2UKGgGR8BXAAAAAAAAaAdLXGgIR0BzaYRqXWvsdX2UKGgGR8BZwAAAAAAAaAdLZ2gIR0BzabqB3A2ydWUu"
70
+ },
71
+ "ep_success_buffer": {
72
+ ":type:": "<class 'collections.deque'>",
73
+ ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
74
+ },
75
+ "_n_updates": 9770,
76
+ "n_steps": 2048,
77
+ "gamma": 0.99,
78
+ "gae_lambda": 0.95,
79
+ "ent_coef": 0.0,
80
+ "vf_coef": 0.5,
81
+ "max_grad_norm": 0.5,
82
+ "batch_size": 64,
83
+ "n_epochs": 10,
84
+ "clip_range": {
85
+ ":type:": "<class 'function'>",
86
+ ":serialized:": "gAWV/wIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwNX2J1aWx0aW5fdHlwZZSTlIwKTGFtYmRhVHlwZZSFlFKUKGgCjAhDb2RlVHlwZZSFlFKUKEsBSwBLAEsBSwFLE0MEiABTAJROhZQpjAFflIWUjGcvVXNlcnMvdGhlYXZlYXNzby9taW5pZm9yZ2UzL2VudnMvaW50cm9STC9saWIvcHl0aG9uMy45L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxnL1VzZXJzL3RoZWF2ZWFzc28vbWluaWZvcmdlMy9lbnZzL2ludHJvUkwvbGliL3B5dGhvbjMuOS9zaXRlLXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGggfZR9lChoF2gOjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoGIwHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/yZmZmZmZmoWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
87
+ },
88
+ "clip_range_vf": null,
89
+ "normalize_advantage": true,
90
+ "target_kl": null
91
+ }
ppo-mountaincar-v0/policy.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45693dd5d0443ac23fb575222c43407d4f9a50538620a4d4bc77f98cc5887735
3
+ size 77981
ppo-mountaincar-v0/policy.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1846b1e796086b58e19bcfaeb84929f3876a79a9d306b4603d3df964e5b5bc2
3
+ size 39745
ppo-mountaincar-v0/pytorch_variables.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d030ad8db708280fcae77d87e973102039acd23a11bdecc3db8eb6c0ac940ee1
3
+ size 431
ppo-mountaincar-v0/system_info.txt ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ OS: macOS-12.3.1-arm64-arm-64bit Darwin Kernel Version 21.4.0: Fri Mar 18 00:46:32 PDT 2022; root:xnu-8020.101.4~15/RELEASE_ARM64_T6000
2
+ Python: 3.9.12
3
+ Stable-Baselines3: 1.5.0
4
+ PyTorch: 1.11.0
5
+ GPU Enabled: False
6
+ Numpy: 1.22.3
7
+ Gym: 0.21.0
replay.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12e55d12c57b547e49fec877ac112d39a27c14e117bb4c337b55717e1afbc54a
3
+ size 477576
results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"mean_reward": -97.2, "std_reward": 7.124605252222751, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-05-06T20:11:15.040017"}