first commit
Browse files- README.md +1 -12
- config.json +1 -1
- ppo-LunarLander-v2.zip +2 -2
- ppo-LunarLander-v2/data +18 -18
- ppo-LunarLander-v2/policy.optimizer.pth +1 -1
- ppo-LunarLander-v2/policy.pth +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
README.md
CHANGED
@@ -16,7 +16,7 @@ model-index:
|
|
16 |
type: LunarLander-v2
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
-
value:
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
@@ -34,15 +34,4 @@ from stable_baselines3 import ...
|
|
34 |
from huggingface_sb3 import load_from_hub
|
35 |
|
36 |
...
|
37 |
-
|
38 |
-
model = PPO('MlpPolicy',
|
39 |
-
env,
|
40 |
-
# batch_size = 128,
|
41 |
-
n_epochs = 7,
|
42 |
-
gamma = 0.995,
|
43 |
-
gae_lambda = 0.97,
|
44 |
-
ent_coef = 0.01,
|
45 |
-
verbose=1,
|
46 |
-
tensorboard_log='tb_logs/lr')
|
47 |
-
|
48 |
```
|
|
|
16 |
type: LunarLander-v2
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
+
value: 283.64 +/- 17.16
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
|
|
34 |
from huggingface_sb3 import load_from_hub
|
35 |
|
36 |
...
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
```
|
config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x797893aa37f0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x797893aa3880>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x797893aa3910>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x797893aa39a0>", "_build": "<function ActorCriticPolicy._build at 0x797893aa3a30>", "forward": "<function ActorCriticPolicy.forward at 0x797893aa3ac0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x797893aa3b50>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x797893aa3be0>", "_predict": "<function ActorCriticPolicy._predict at 0x797893aa3c70>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x797893aa3d00>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x797893aa3d90>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x797893aa3e20>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x797893a49780>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 3014656, "_total_timesteps": 3000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1706967059420466826, "learning_rate": 0.0003, "tensorboard_log": "tb_logs/lr", "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAM3SLDz23Be6t9iGORNmITWR4lU7mvScuAAAgD8AAIA/gCw3vY+iebqa8049FXIss+7qEzuS4GSzAACAPwAAgD8zKIY8xLO3P2Ln0z23sge+UYADPTJ83zwAAAAAAAAAALMnJr2uZa66xkcqs2tGL7D3DDK6GxTOMwAAgD8AAIA/ZgVjvXts1ro5/549PtK6OovooLtzsqI7AACAPwAAgD/mDai9H823udN6Fz7Xqpu88v04u9ZDiL0AAAAAAACAP+ZgR72p2wa8Hrw3Posoxb1+4ke86m3pvgAAgD8AAIA/c984vkPcXz9wYou+Wpwvv0dJrr5cWiS+AAAAAAAAAADN5EA9KWQrui6g3rYzaNixAaZ1O3dQBTYAAIA/AACAP/Mykr0iLYk/uJBlvjWVYr+tCwy+lVG+vQAAAAAAAAAAM80pvOFAg7r89SezDQ/Nqu1MVzvJeMozAACAPwAAgD9r/KC+/PTxPmIVdT4SjiK/nJqJvkFHJj4AAAAAAAAAAACJfb3XAwa7r2LEPPDUeTyqbx28i4dZPQAAAAAAAIA/AFfePKQlfLt6ONW9jbklPe9dpTyWaAq+AACAPwAAgD+aeaa85kOnP0pGRr5gThW/N7AevRrTzL0AAAAAAAAAAIDdBj24xvG5ZuAIvWXPA7Z/ZKw7qvRyNQAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.004885333333333408, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHF4EWEbo8qMAWyUS7WMAXSUR0DC2Y3Ru0kXdX2UKGgGR0BzQF+KCQLeaAdLrWgIR0DC2Z7fP5YYdX2UKGgGR0Bw13TI/7iyaAdLs2gIR0DC2aQNutOmdX2UKGgGR0BxUr5xiobXaAdLpmgIR0DC2bE6zVtodX2UKGgGR0Bw4JcHGCI2aAdLtmgIR0DC2bEiD/VBdX2UKGgGR0By8CQcPvroaAdLy2gIR0DC2b5hz/6wdX2UKGgGR0BxmgEyLyc1aAdLwWgIR0DC2csbNr0rdX2UKGgGR0BKI8RlHz6KaAdLdWgIR0DC2d9ZaFEidX2UKGgGR0BxIBTfixVyaAdLp2gIR0DC2fE6eXiSdX2UKGgGR0ByXYNz8xbjaAdLtWgIR0DC2fIUzsQedX2UKGgGR0Bw2s8TzundaAdLnmgIR0DC2fphWo3rdX2UKGgGR0ByLKthd+ocaAdLpmgIR0DC2ft5GBnSdX2UKGgGR0Bxg9W7voeQaAdLs2gIR0DC2gJHG0eEdX2UKGgGR0BzlOIP9UCJaAdLvWgIR0DC2iYClrM1dX2UKGgGR0BuwwnMMZxaaAdLmmgIR0DC2jpQgs9TdX2UKGgGR0BySrTQVsUJaAdLs2gIR0DC2juIj4YadX2UKGgGR0Bxv05o4+8oaAdLumgIR0DC2j0ifQKKdX2UKGgGR0BzhAxrSE13aAdLwWgIR0DC2jxDZ13ddX2UKGgGR0BylMXhwVCYaAdLmWgIR0DC2kS2a2F4dX2UKGgGR0Bu3k3XI2fkaAdLrGgIR0DC2lAhY/3WdX2UKGgGR0ByRuumrKeTaAdLomgIR0DC2lHzJ6ppdX2UKGgGR0ByiEK9f1HwaAdLy2gIR0DC2ld74SHudX2UKGgGR0BxwICPp6hQaAdLvGgIR0DC2mvZ/Tb4dX2UKGgGR0BxEnDAJswdaAdLlmgIR0DC2mwK+i8GdX2UKGgGR0Bwgjr9l2/0aAdLqmgIR0DC2m53os7NdX2UKGgGR0BzC2sDGLk0aAdLqWgIR0DC2n+p84PxdX2UKGgGR0BzN0bvPTodaAdLt2gIR0DC2oM5EMLGdX2UKGgGR0BxlVs0pEx7aAdLuGgIR0DC2ojsa86FdX2UKGgGR0BxwwJ0GNaRaAdLuWgIR0DC2o4H9m6HdX2UKGgGR0BxEjiXIEKWaAdLhmgIR0DC2px9b5dodX2UKGgGR0Bw3i7z06HTaAdLmGgIR0DC2p+EVWS2dX2UKGgGR0BxXDQiRnvlaAdLx2gIR0DC2qx8lXzUdX2UKGgGR0ByqHFUADJVaAdLw2gIR0DC2sBOHnEEdX2UKGgGR0BywsNH6MzeaAdLy2gIR0DC2sTlq8DkdX2UKGgGR0BxBz83uNPyaAdLoWgIR0DC2sTuOS4fdX2UKGgGR0B0CFpyp71JaAdLuGgIR0DC2sy/Zdv9dX2UKGgGR0BzX/WQOnVHaAdL5GgIR0DC2tW6NEPUdX2UKGgGR0BzEWGYa5wwaAdLz2gIR0DC2t3C9AX3dX2UKGgGR0ByLp8zAN5MaAdLtWgIR0DC2ult2s7udX2UKGgGR0ByRf9pAUtaaAdLuGgIR0DC2ulFUhmodX2UKGgGR0Bv/3C9AX2vaAdLmmgIR0DC2usYKpkxdX2UKGgGR0By2nfrKNhmaAdLxWgIR0DC2vGuNgjRdX2UKGgGR0BzUbBEa2nbaAdLqmgIR0DC2vI0ygwodX2UKGgGR0ByK5nEl3QlaAdLsWgIR0DC2v66+WWydX2UKGgGR0Bx33UZvUBoaAdLmWgIR0DC2wbQmeDndX2UKGgGR0BP7hNdqtYCaAdLbmgIR0DC2w0ejmCAdX2UKGgGR0By+MoXsPataAdLvWgIR0DC2wwplSTAdX2UKGgGR0BwmTc6/7BPaAdLqGgIR0DC2w4B3iaRdX2UKGgGR0Byir71qWTpaAdLw2gIR0DC2y/LidaudX2UKGgGR0Bzt0WTHKfWaAdLs2gIR0DC2zeRigCfdX2UKGgGR0ByeKT0QK8daAdLlGgIR0DC2zfsqrimdX2UKGgGR0BwWvGLk0aZaAdLjGgIR0DC20ZkkKNRdX2UKGgGR0ByTGsV+I/JaAdLvWgIR0DC20ucnVoYdX2UKGgGR0Bxm2qXF98aaAdLrmgIR0DC21Ltw71adX2UKGgGR0BzRejxkNF0aAdL2WgIR0DC21bLjghsdX2UKGgGR0ByNGzNUwSKaAdLq2gIR0DC212AAhjfdX2UKGgGR0BzejCQ9zOpaAdLtmgIR0DC225DJEH/dX2UKGgGR0Bx6ck3S8aoaAdLt2gIR0DC22/ICEHudX2UKGgGR0ByeDAWSEDhaAdLzGgIR0DC23VsDW9UdX2UKGgGR0Bv4kBdUsFuaAdLmmgIR0DC23mhdt2tdX2UKGgGR0BzbD8XN1QqaAdLpWgIR0DC23q/XXiBdX2UKGgGR0BwmT0Zm7J5aAdLuWgIR0DC239snAqNdX2UKGgGR0BxsR5prULEaAdLs2gIR0DC24irtE5RdX2UKGgGR0BxQBVAAyVOaAdLsmgIR0DC24nZ7HAAdX2UKGgGR0BxW685CF9KaAdLpmgIR0DC26PDR+jNdX2UKGgGR0Bys4jcEeQuaAdLj2gIR0DC268Mw1zidX2UKGgGR0BxXG8jAzpHaAdLsmgIR0DC27SpNsWPdX2UKGgGR0BzjO2+fywwaAdLumgIR0DC27rwtrbhdX2UKGgGR0BxLrRUm2LHaAdLymgIR0DC29TzRQaadX2UKGgGR0BzCuDZlFtsaAdLvmgIR0DC29ksJ6Y3dX2UKGgGR0BxG0qLCN0eaAdLwGgIR0DC295YDDCQdX2UKGgGR0BxK0l0HQhPaAdLmGgIR0DC2+Qqd6LPdX2UKGgGR0Bwk1ygf2boaAdLqmgIR0DC2+TQu27WdX2UKGgGR0Bxv8snRb8naAdLwmgIR0DC2+bMaCL/dX2UKGgGR0BzZPYh+vyLaAdLuGgIR0DC2/BFPSDzdX2UKGgGR0BylsSCe2/jaAdLuGgIR0DC2/WwTufFdX2UKGgGR0BxAehysCDFaAdLq2gIR0DC2/dbNbC8dX2UKGgGR0BxaJ6Ww/xEaAdLvWgIR0DC2/z+BH09dX2UKGgGR0Bxz4fuCwr2aAdLsmgIR0DC3AbF2mpEdX2UKGgGR0BzmZX+2mYTaAdL3GgIR0DC3CI2dd3TdX2UKGgGR0ByBz/JeVs2aAdLxGgIR0DC3DjMaCL/dX2UKGgGR0BzRa5y2hIwaAdL12gIR0DC3DtSuQp4dX2UKGgGR0BxwyvxH5JsaAdLomgIR0DC3EW03Ov/dX2UKGgGR0Bz7fVhCtzTaAdL22gIR0DC3E2yquKXdX2UKGgGR0BzW4VKwpvxaAdL2GgIR0DC3FEFt8/mdX2UKGgGR0ByCr3wkPc0aAdLsmgIR0DC3FOi8FpxdX2UKGgGR0BxLwx59mYjaAdLp2gIR0DC3FdfJFLGdX2UKGgGR0Bzzcy+HrQgaAdLsWgIR0DC3FfGS6lMdX2UKGgGR0Bw7JkTYdyUaAdLmGgIR0DC3FyakRBedX2UKGgGR0Bz54fHPu5SaAdLtmgIR0DC3F/8KohqdX2UKGgGR0BxoIg/1QIlaAdLsmgIR0DC3F/Y150KdX2UKGgGR0BwiaKsMiKSaAdLrWgIR0DC3GPJJXhgdX2UKGgGR0BwgwtsenyeaAdLpWgIR0DC3GUleF+NdX2UKGgGR0ByP1i+cpb2aAdLoWgIR0DC3GgBHTZydX2UKGgGR0BvdS7Ackt3aAdLkGgIR0DC3H8AvL5idX2UKGgGR0BxbcYgq3EyaAdLxGgIR0DC3IaCFsYVdX2UKGgGR0BxlmtU4rBkaAdLq2gIR0DC3K3np0OmdX2UKGgGR0B0OHYAbQ1KaAdLtGgIR0DC3LLZ+QU6dX2UKGgGR0ByuNvsJIDpaAdLpmgIR0DC3L4CyQgcdX2UKGgGR0BwpVd6cAinaAdLpWgIR0DC3MG5lOGkdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 644, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 2048, "gamma": 0.995, "gae_lambda": 0.97, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 7, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu121", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7892da9b2ef0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7892da9b2f80>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7892da9b3010>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7892da9b30a0>", "_build": "<function ActorCriticPolicy._build at 0x7892da9b3130>", "forward": "<function ActorCriticPolicy.forward at 0x7892da9b31c0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7892da9b3250>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7892da9b32e0>", "_predict": "<function ActorCriticPolicy._predict at 0x7892da9b3370>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7892da9b3400>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7892da9b3490>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7892da9b3520>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7892da94fbc0>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 3014656, "_total_timesteps": 3000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1707070109434399459, "learning_rate": 0.0003, "tensorboard_log": "tb_logs/lr", "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGbuErxcj1u63smXs9MI6C5pQgA7D2CyMwAAgD8AAIA/7WAZPp2LAD9OQ4y9Qmoov+DqNz5I04i+AAAAAAAAAAAAKJ47cMq0P4hK+j7uO4k9cQm3u1jH4r0AAAAAAAAAAJrTXLz2eHW6gjhlNzzJLDE5cXq7rJeEtgAAgD8AAIA/mta2vMhpiT8q1nO91mxlv7pXorwY5+E7AAAAAAAAAADDDE2+duVFP04RB76PsQq/ND3YvnIe9b0AAAAAAAAAAJoAJr3TwTM/PqOVvQybhr+JABW9DlQKPAAAAAAAAAAAmk3OvZKNtz9Og5y+BmW1vtouyr0z/zC+AAAAAAAAAADNXii+RTKqP4SMEr/6/vC+5mSSvmboBr8AAAAAAAAAAGb1gb1PNHa8osQAPr7zaz30CJu90ky7PAAAgD8AAIA/AKYevYWL9jifxx++BfkaM1xqzjuez3yzAACAPwAAgD9GHjW+gQYLP3EPPj7WxyO/6jeyvs9Msz4AAAAAAAAAAKCeFT6eYtI+oqlFvkdi/74359A9gNVSvgAAAAAAAAAAMy6jPVG84D0HM6q+vVr0vr/B3b2BzTC+AAAAAAAAAADA/Ne9qXhEPWK/vj3BXsm+sNxRuxgn0jsAAAAAAAAAAPOvcj7o7JY/8qgKP2RTBL+hsgQ/pLqVPgAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAQAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.004885333333333408, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHA25OvdM0yMAWyUS5eMAXSUR0DD40h/oaDPdX2UKGgGR0Bxu4slLOAzaAdLp2gIR0DD405g5R0mdX2UKGgGR0BxxPtVrAP/aAdLt2gIR0DD42uI2wV1dX2UKGgGR0BxlOXiR4hVaAdLmmgIR0DD43sPUaybdX2UKGgGR0BxA9utOmBOaAdLrWgIR0DD44TQPZqVdX2UKGgGR0BwlXqD9OynaAdLrmgIR0DD46IGGEf1dX2UKGgGR0Bx4eIWP91maAdLiGgIR0DD46j6DXe4dX2UKGgGR0ByOVUYKpkxaAdLu2gIR0DD465tHhCMdX2UKGgGR0Bze9Au7HyVaAdLtWgIR0DD47J3kgfVdX2UKGgGR0BxJj69CeEqaAdLq2gIR0DD47t1GLDRdX2UKGgGR0BwDXq9oN/faAdLimgIR0DD48Dkhib2dX2UKGgGR0ByL57+kxh2aAdLj2gIR0DD48JOnEVGdX2UKGgGR0ByLe9g4OtoaAdLsmgIR0DD49H6uW8idX2UKGgGR0BwsNog3cYZaAdLp2gIR0DD49HLX+VDdX2UKGgGR0ByENdUsFt9aAdLp2gIR0DD492VxCIDdX2UKGgGR0Bw+uBtk4FSaAdLrmgIR0DD4+SnvUjLdX2UKGgGR0BBCmKAJ9iMaAdLWGgIR0DD4/q6Ymb9dX2UKGgGR0BxdGI0qH45aAdLwGgIR0DD4/10vGp/dX2UKGgGR0BwQGqjrRjSaAdLoWgIR0DD5AH1BdD6dX2UKGgGR0ByABNahYeUaAdL1WgIR0DD5AsYXO4YdX2UKGgGR0BygoUg0TDgaAdLvWgIR0DD5C/0dzXCdX2UKGgGR0Bxguojv/ipaAdLnWgIR0DD5DlrdnCgdX2UKGgGR0BwKwOvt+kQaAdLvWgIR0DD5DsI1LrYdX2UKGgGR0ByP3JRwZO0aAdLgmgIR0DD5D8abWmQdX2UKGgGR0BwPUtoSL62aAdLjGgIR0DD5EN9roGIdX2UKGgGR0BxdIJLM9r5aAdLomgIR0DD5EtXDFZQdX2UKGgGR0Bw/Ejnmq5taAdLoWgIR0DD5Fd1SwW4dX2UKGgGR0BzNod7v5P/aAdLuGgIR0DD5Fu938oAdX2UKGgGR0By3jNqxkd4aAdLqmgIR0DD5GddPci4dX2UKGgGR0BxQMAsCkoGaAdLtmgIR0DD5G8pG4I9dX2UKGgGR0BxWhtuUD+zaAdLq2gIR0DD5HYDFId3dX2UKGgGR0ByQuSntOVPaAdLqmgIR0DD5IVd/rjYdX2UKGgGR0ByISy+pOvdaAdLzmgIR0DD5IdpoK2KdX2UKGgGR0ByCAnqmj0uaAdLq2gIR0DD5Ikm8dxRdX2UKGgGR0BywmvdM0xeaAdLyGgIR0DD5Jd5nlGPdX2UKGgGR0BybCICU5dXaAdLuWgIR0DD5JmZE2HddX2UKGgGR0Bx+b63y7PIaAdLlmgIR0DD5KMJY1YRdX2UKGgGR0BuCAjUutfYaAdLqmgIR0DD5KYnlXA/dX2UKGgGR0Bxh96jWTX8aAdLpGgIR0DD5KfW6K+BdX2UKGgGR0Bx45M6BAfMaAdLtmgIR0DD5LRu4wyqdX2UKGgGR0BwzzGff4yoaAdLl2gIR0DD5LfYao/BdX2UKGgGR0Bzbd3W4EwGaAdLumgIR0DD5LyAOJ+EdX2UKGgGR0Bx3lQLux8laAdLt2gIR0DD5L+7QLNOdX2UKGgGR0Bu/NyT6i0waAdLpGgIR0DD5NGDnNgSdX2UKGgGR0BzrALkS26TaAdLt2gIR0DD5NGDJ2dNdX2UKGgGR0BxEeqZML4OaAdLsmgIR0DD5OKpT/ACdX2UKGgGR0BxMBLdvbXZaAdLsGgIR0DD5OiXyAhCdX2UKGgGR0BxDldIGyHEaAdLqWgIR0DD5PZ2GIsRdX2UKGgGR0Bx1VWo3rD7aAdLmWgIR0DD5PxRbbDedX2UKGgGR0Bw0C/i5uqFaAdLtGgIR0DD5P/kHUtqdX2UKGgGR0BzRSyZ8a4uaAdLvmgIR0DD5QLURWcSdX2UKGgGR0BxGhnuiN83aAdLpmgIR0DD5RAT7EYPdX2UKGgGR0ByxQIY3vQXaAdLtmgIR0DD5RCv9tMxdX2UKGgGR0BwLbOgQHzIaAdLo2gIR0DD5SETSLIgdX2UKGgGR0ByH7aTOgQIaAdLvWgIR0DD5SVYlpoLdX2UKGgGR0BzPNZ/0/W2aAdLxWgIR0DD5SipHZsbdX2UKGgGR0BxSftkWhysaAdLtGgIR0DD5S9Q9A5adX2UKGgGR0Bwd3wVj7Q+aAdLr2gIR0DD5TP6AOJ+dX2UKGgGR0Bz3yyjYZl4aAdLw2gIR0DD5T11+y7gdX2UKGgGR0BxkfDuSfUXaAdLumgIR0DD5UvJcPe6dX2UKGgGR0BzLgu3+dbxaAdLwWgIR0DD5VBftx+8dX2UKGgGR0BwEla8pTddaAdLlGgIR0DD5ValJpWWdX2UKGgGR0BxmFsxfv4NaAdLu2gIR0DD5V0bo8p1dX2UKGgGR0By9m53C9AYaAdLpWgIR0DD5WlIZqEfdX2UKGgGR0BzQ0Tj/+85aAdLzmgIR0DD5W5gTh5xdX2UKGgGR0Byy9XU6PsBaAdLqWgIR0DD5W8sOG0vdX2UKGgGR0BzRB55Z8rqaAdLu2gIR0DD5XScAimmdX2UKGgGR0BuxsIeHSF5aAdLn2gIR0DD5XaR8twrdX2UKGgGR0Bzc7A57w8XaAdLrmgIR0DD5Y9wiqyXdX2UKGgGR0By+pcs189faAdLzmgIR0DD5ZSmCROldX2UKGgGR0Bze1SP2f03aAdLuGgIR0DD5Z8052hadX2UKGgGR0By+PW6K+BZaAdLymgIR0DD5afOObRXdX2UKGgGR0Bxlcf0VafSaAdLtmgIR0DD5aoAIY3vdX2UKGgGR0BzDaxdIGyHaAdLxWgIR0DD5a6vC/GmdX2UKGgGR0BwsM4cWCVbaAdLm2gIR0DD5bUFQl8gdX2UKGgGR0BzWTcxj8UFaAdLxWgIR0DD5b0Yht+DdX2UKGgGR0Byp7DO1OTJaAdLsWgIR0DD5cngYP5IdX2UKGgGR0ByXV5C4SYgaAdLsWgIR0DD5dCDRMN+dX2UKGgGR0Bzv7c32mHhaAdL2mgIR0DD5dj3sXzldX2UKGgGR0BxZ5FKCg9NaAdLnGgIR0DD5drvNNahdX2UKGgGR0Bwp7q9oN/faAdLpWgIR0DD5dt36hxpdX2UKGgGR0BzUAKiO/+LaAdLsGgIR0DD5d0o4MnadX2UKGgGR0Bz3kAbQ1JlaAdLrWgIR0DD5d/qZ+hHdX2UKGgGR0By/xXPqs2faAdLnWgIR0DD5fUjVx0ddX2UKGgGR0BzM1rvb48EaAdLxWgIR0DD5fXu7YkFdX2UKGgGR0ByDhptaY/naAdLpWgIR0DD5f8Vgx8EdX2UKGgGR0BwZoqQRwqBaAdLqGgIR0DD5gqQRwqBdX2UKGgGR0BxIPevZAY6aAdLnmgIR0DD5g6/bj95dX2UKGgGR0BwCS0x/NJOaAdLn2gIR0DD5hR2KVIJdX2UKGgGR0Bw5XHo5ggHaAdLr2gIR0DD5heq5sj3dX2UKGgGR0Bx99Pacqe9aAdLuWgIR0DD5ixbD/EPdX2UKGgGR0Bx2cnJDE3saAdLjGgIR0DD5jfIlt0ndX2UKGgGR0BzQXzVc2R8aAdLwmgIR0DD5jqXIEKWdX2UKGgGR0BzIMJLM9r5aAdLvGgIR0DD5kp9y926dX2UKGgGR0Bx22Qmu1WsaAdLrWgIR0DD5ktOEdvLdX2UKGgGR0Bw7xsUIsy0aAdLqWgIR0DD5k2zF+/hdX2UKGgGR0Bx+O+bmU4aaAdLtGgIR0DD5k3vc8DCdX2UKGgGR0BxHPzND+iraAdLtmgIR0DD5lFFQVKxdX2UKGgGR0BzJ47GNrCWaAdL1mgIR0DD5lRX+2mYdX2UKGgGR0ByuGvt+kP+aAdLqGgIR0DD5mKhcqvvdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 644, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 2048, "gamma": 0.995, "gae_lambda": 0.97, "ent_coef": 0.005, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 7, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu121", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
ppo-LunarLander-v2.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:145d99ec5f8177a8dbe4955c46e66101a6488967d2b98ae94a5384bb0f8b66d9
|
3 |
+
size 147945
|
ppo-LunarLander-v2/data
CHANGED
@@ -4,20 +4,20 @@
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
11 |
-
"_build": "<function ActorCriticPolicy._build at
|
12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
-
"_abc_impl": "<_abc._abc_data object at
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {},
|
@@ -26,16 +26,16 @@
|
|
26 |
"_num_timesteps_at_start": 0,
|
27 |
"seed": null,
|
28 |
"action_noise": null,
|
29 |
-
"start_time":
|
30 |
"learning_rate": 0.0003,
|
31 |
"tensorboard_log": "tb_logs/lr",
|
32 |
"_last_obs": {
|
33 |
":type:": "<class 'numpy.ndarray'>",
|
34 |
-
":serialized:": "
|
35 |
},
|
36 |
"_last_episode_starts": {
|
37 |
":type:": "<class 'numpy.ndarray'>",
|
38 |
-
":serialized:": "
|
39 |
},
|
40 |
"_last_original_obs": null,
|
41 |
"_episode_num": 0,
|
@@ -45,7 +45,7 @@
|
|
45 |
"_stats_window_size": 100,
|
46 |
"ep_info_buffer": {
|
47 |
":type:": "<class 'collections.deque'>",
|
48 |
-
":serialized:": "
|
49 |
},
|
50 |
"ep_success_buffer": {
|
51 |
":type:": "<class 'collections.deque'>",
|
@@ -80,7 +80,7 @@
|
|
80 |
"n_steps": 2048,
|
81 |
"gamma": 0.995,
|
82 |
"gae_lambda": 0.97,
|
83 |
-
"ent_coef": 0.
|
84 |
"vf_coef": 0.5,
|
85 |
"max_grad_norm": 0.5,
|
86 |
"batch_size": 64,
|
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7892da9b2ef0>",
|
8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7892da9b2f80>",
|
9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7892da9b3010>",
|
10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7892da9b30a0>",
|
11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7892da9b3130>",
|
12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7892da9b31c0>",
|
13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7892da9b3250>",
|
14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7892da9b32e0>",
|
15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7892da9b3370>",
|
16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7892da9b3400>",
|
17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7892da9b3490>",
|
18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7892da9b3520>",
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
+
"_abc_impl": "<_abc._abc_data object at 0x7892da94fbc0>"
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {},
|
|
|
26 |
"_num_timesteps_at_start": 0,
|
27 |
"seed": null,
|
28 |
"action_noise": null,
|
29 |
+
"start_time": 1707070109434399459,
|
30 |
"learning_rate": 0.0003,
|
31 |
"tensorboard_log": "tb_logs/lr",
|
32 |
"_last_obs": {
|
33 |
":type:": "<class 'numpy.ndarray'>",
|
34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGbuErxcj1u63smXs9MI6C5pQgA7D2CyMwAAgD8AAIA/7WAZPp2LAD9OQ4y9Qmoov+DqNz5I04i+AAAAAAAAAAAAKJ47cMq0P4hK+j7uO4k9cQm3u1jH4r0AAAAAAAAAAJrTXLz2eHW6gjhlNzzJLDE5cXq7rJeEtgAAgD8AAIA/mta2vMhpiT8q1nO91mxlv7pXorwY5+E7AAAAAAAAAADDDE2+duVFP04RB76PsQq/ND3YvnIe9b0AAAAAAAAAAJoAJr3TwTM/PqOVvQybhr+JABW9DlQKPAAAAAAAAAAAmk3OvZKNtz9Og5y+BmW1vtouyr0z/zC+AAAAAAAAAADNXii+RTKqP4SMEr/6/vC+5mSSvmboBr8AAAAAAAAAAGb1gb1PNHa8osQAPr7zaz30CJu90ky7PAAAgD8AAIA/AKYevYWL9jifxx++BfkaM1xqzjuez3yzAACAPwAAgD9GHjW+gQYLP3EPPj7WxyO/6jeyvs9Msz4AAAAAAAAAAKCeFT6eYtI+oqlFvkdi/74359A9gNVSvgAAAAAAAAAAMy6jPVG84D0HM6q+vVr0vr/B3b2BzTC+AAAAAAAAAADA/Ne9qXhEPWK/vj3BXsm+sNxRuxgn0jsAAAAAAAAAAPOvcj7o7JY/8qgKP2RTBL+hsgQ/pLqVPgAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
35 |
},
|
36 |
"_last_episode_starts": {
|
37 |
":type:": "<class 'numpy.ndarray'>",
|
38 |
+
":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAQAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="
|
39 |
},
|
40 |
"_last_original_obs": null,
|
41 |
"_episode_num": 0,
|
|
|
45 |
"_stats_window_size": 100,
|
46 |
"ep_info_buffer": {
|
47 |
":type:": "<class 'collections.deque'>",
|
48 |
+
":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHA25OvdM0yMAWyUS5eMAXSUR0DD40h/oaDPdX2UKGgGR0Bxu4slLOAzaAdLp2gIR0DD405g5R0mdX2UKGgGR0BxxPtVrAP/aAdLt2gIR0DD42uI2wV1dX2UKGgGR0BxlOXiR4hVaAdLmmgIR0DD43sPUaybdX2UKGgGR0BxA9utOmBOaAdLrWgIR0DD44TQPZqVdX2UKGgGR0BwlXqD9OynaAdLrmgIR0DD46IGGEf1dX2UKGgGR0Bx4eIWP91maAdLiGgIR0DD46j6DXe4dX2UKGgGR0ByOVUYKpkxaAdLu2gIR0DD465tHhCMdX2UKGgGR0Bze9Au7HyVaAdLtWgIR0DD47J3kgfVdX2UKGgGR0BxJj69CeEqaAdLq2gIR0DD47t1GLDRdX2UKGgGR0BwDXq9oN/faAdLimgIR0DD48Dkhib2dX2UKGgGR0ByL57+kxh2aAdLj2gIR0DD48JOnEVGdX2UKGgGR0ByLe9g4OtoaAdLsmgIR0DD49H6uW8idX2UKGgGR0BwsNog3cYZaAdLp2gIR0DD49HLX+VDdX2UKGgGR0ByENdUsFt9aAdLp2gIR0DD492VxCIDdX2UKGgGR0Bw+uBtk4FSaAdLrmgIR0DD4+SnvUjLdX2UKGgGR0BBCmKAJ9iMaAdLWGgIR0DD4/q6Ymb9dX2UKGgGR0BxdGI0qH45aAdLwGgIR0DD4/10vGp/dX2UKGgGR0BwQGqjrRjSaAdLoWgIR0DD5AH1BdD6dX2UKGgGR0ByABNahYeUaAdL1WgIR0DD5AsYXO4YdX2UKGgGR0BygoUg0TDgaAdLvWgIR0DD5C/0dzXCdX2UKGgGR0Bxguojv/ipaAdLnWgIR0DD5DlrdnCgdX2UKGgGR0BwKwOvt+kQaAdLvWgIR0DD5DsI1LrYdX2UKGgGR0ByP3JRwZO0aAdLgmgIR0DD5D8abWmQdX2UKGgGR0BwPUtoSL62aAdLjGgIR0DD5EN9roGIdX2UKGgGR0BxdIJLM9r5aAdLomgIR0DD5EtXDFZQdX2UKGgGR0Bw/Ejnmq5taAdLoWgIR0DD5Fd1SwW4dX2UKGgGR0BzNod7v5P/aAdLuGgIR0DD5Fu938oAdX2UKGgGR0By3jNqxkd4aAdLqmgIR0DD5GddPci4dX2UKGgGR0BxQMAsCkoGaAdLtmgIR0DD5G8pG4I9dX2UKGgGR0BxWhtuUD+zaAdLq2gIR0DD5HYDFId3dX2UKGgGR0ByQuSntOVPaAdLqmgIR0DD5IVd/rjYdX2UKGgGR0ByISy+pOvdaAdLzmgIR0DD5IdpoK2KdX2UKGgGR0ByCAnqmj0uaAdLq2gIR0DD5Ikm8dxRdX2UKGgGR0BywmvdM0xeaAdLyGgIR0DD5Jd5nlGPdX2UKGgGR0BybCICU5dXaAdLuWgIR0DD5JmZE2HddX2UKGgGR0Bx+b63y7PIaAdLlmgIR0DD5KMJY1YRdX2UKGgGR0BuCAjUutfYaAdLqmgIR0DD5KYnlXA/dX2UKGgGR0Bxh96jWTX8aAdLpGgIR0DD5KfW6K+BdX2UKGgGR0Bx45M6BAfMaAdLtmgIR0DD5LRu4wyqdX2UKGgGR0BwzzGff4yoaAdLl2gIR0DD5LfYao/BdX2UKGgGR0Bzbd3W4EwGaAdLumgIR0DD5LyAOJ+EdX2UKGgGR0Bx3lQLux8laAdLt2gIR0DD5L+7QLNOdX2UKGgGR0Bu/NyT6i0waAdLpGgIR0DD5NGDnNgSdX2UKGgGR0BzrALkS26TaAdLt2gIR0DD5NGDJ2dNdX2UKGgGR0BxEeqZML4OaAdLsmgIR0DD5OKpT/ACdX2UKGgGR0BxMBLdvbXZaAdLsGgIR0DD5OiXyAhCdX2UKGgGR0BxDldIGyHEaAdLqWgIR0DD5PZ2GIsRdX2UKGgGR0Bx1VWo3rD7aAdLmWgIR0DD5PxRbbDedX2UKGgGR0Bw0C/i5uqFaAdLtGgIR0DD5P/kHUtqdX2UKGgGR0BzRSyZ8a4uaAdLvmgIR0DD5QLURWcSdX2UKGgGR0BxGhnuiN83aAdLpmgIR0DD5RAT7EYPdX2UKGgGR0ByxQIY3vQXaAdLtmgIR0DD5RCv9tMxdX2UKGgGR0BwLbOgQHzIaAdLo2gIR0DD5SETSLIgdX2UKGgGR0ByH7aTOgQIaAdLvWgIR0DD5SVYlpoLdX2UKGgGR0BzPNZ/0/W2aAdLxWgIR0DD5SipHZsbdX2UKGgGR0BxSftkWhysaAdLtGgIR0DD5S9Q9A5adX2UKGgGR0Bwd3wVj7Q+aAdLr2gIR0DD5TP6AOJ+dX2UKGgGR0Bz3yyjYZl4aAdLw2gIR0DD5T11+y7gdX2UKGgGR0BxkfDuSfUXaAdLumgIR0DD5UvJcPe6dX2UKGgGR0BzLgu3+dbxaAdLwWgIR0DD5VBftx+8dX2UKGgGR0BwEla8pTddaAdLlGgIR0DD5ValJpWWdX2UKGgGR0BxmFsxfv4NaAdLu2gIR0DD5V0bo8p1dX2UKGgGR0By9m53C9AYaAdLpWgIR0DD5WlIZqEfdX2UKGgGR0BzQ0Tj/+85aAdLzmgIR0DD5W5gTh5xdX2UKGgGR0Byy9XU6PsBaAdLqWgIR0DD5W8sOG0vdX2UKGgGR0BzRB55Z8rqaAdLu2gIR0DD5XScAimmdX2UKGgGR0BuxsIeHSF5aAdLn2gIR0DD5XaR8twrdX2UKGgGR0Bzc7A57w8XaAdLrmgIR0DD5Y9wiqyXdX2UKGgGR0By+pcs189faAdLzmgIR0DD5ZSmCROldX2UKGgGR0Bze1SP2f03aAdLuGgIR0DD5Z8052hadX2UKGgGR0By+PW6K+BZaAdLymgIR0DD5afOObRXdX2UKGgGR0Bxlcf0VafSaAdLtmgIR0DD5aoAIY3vdX2UKGgGR0BzDaxdIGyHaAdLxWgIR0DD5a6vC/GmdX2UKGgGR0BwsM4cWCVbaAdLm2gIR0DD5bUFQl8gdX2UKGgGR0BzWTcxj8UFaAdLxWgIR0DD5b0Yht+DdX2UKGgGR0Byp7DO1OTJaAdLsWgIR0DD5cngYP5IdX2UKGgGR0ByXV5C4SYgaAdLsWgIR0DD5dCDRMN+dX2UKGgGR0Bzv7c32mHhaAdL2mgIR0DD5dj3sXzldX2UKGgGR0BxZ5FKCg9NaAdLnGgIR0DD5drvNNahdX2UKGgGR0Bwp7q9oN/faAdLpWgIR0DD5dt36hxpdX2UKGgGR0BzUAKiO/+LaAdLsGgIR0DD5d0o4MnadX2UKGgGR0Bz3kAbQ1JlaAdLrWgIR0DD5d/qZ+hHdX2UKGgGR0By/xXPqs2faAdLnWgIR0DD5fUjVx0ddX2UKGgGR0BzM1rvb48EaAdLxWgIR0DD5fXu7YkFdX2UKGgGR0ByDhptaY/naAdLpWgIR0DD5f8Vgx8EdX2UKGgGR0BwZoqQRwqBaAdLqGgIR0DD5gqQRwqBdX2UKGgGR0BxIPevZAY6aAdLnmgIR0DD5g6/bj95dX2UKGgGR0BwCS0x/NJOaAdLn2gIR0DD5hR2KVIJdX2UKGgGR0Bw5XHo5ggHaAdLr2gIR0DD5heq5sj3dX2UKGgGR0Bx99Pacqe9aAdLuWgIR0DD5ixbD/EPdX2UKGgGR0Bx2cnJDE3saAdLjGgIR0DD5jfIlt0ndX2UKGgGR0BzQXzVc2R8aAdLwmgIR0DD5jqXIEKWdX2UKGgGR0BzIMJLM9r5aAdLvGgIR0DD5kp9y926dX2UKGgGR0Bx22Qmu1WsaAdLrWgIR0DD5ktOEdvLdX2UKGgGR0Bw7xsUIsy0aAdLqWgIR0DD5k2zF+/hdX2UKGgGR0Bx+O+bmU4aaAdLtGgIR0DD5k3vc8DCdX2UKGgGR0BxHPzND+iraAdLtmgIR0DD5lFFQVKxdX2UKGgGR0BzJ47GNrCWaAdL1mgIR0DD5lRX+2mYdX2UKGgGR0ByuGvt+kP+aAdLqGgIR0DD5mKhcqvvdWUu"
|
49 |
},
|
50 |
"ep_success_buffer": {
|
51 |
":type:": "<class 'collections.deque'>",
|
|
|
80 |
"n_steps": 2048,
|
81 |
"gamma": 0.995,
|
82 |
"gae_lambda": 0.97,
|
83 |
+
"ent_coef": 0.005,
|
84 |
"vf_coef": 0.5,
|
85 |
"max_grad_norm": 0.5,
|
86 |
"batch_size": 64,
|
ppo-LunarLander-v2/policy.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 88362
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e200d407e62474ec1ef1065fa7ea2d61a7f1b7f2a9ffd5493f19956c8cd1212
|
3 |
size 88362
|
ppo-LunarLander-v2/policy.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 43762
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ef275127e274a61ebf32c4337f84f12df0d364c7ff73db73b24c72fb57a3864
|
3 |
size 43762
|
replay.mp4
CHANGED
Binary files a/replay.mp4 and b/replay.mp4 differ
|
|
results.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"mean_reward":
|
|
|
1 |
+
{"mean_reward": 283.6352223, "std_reward": 17.163272812998546, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-02-04T19:30:54.304061"}
|