Initial commit
Browse files- .gitattributes +1 -0
- README.md +1 -1
- a2c-AntBulletEnv-v0.zip +2 -2
- a2c-AntBulletEnv-v0/data +21 -21
- a2c-AntBulletEnv-v0/policy.optimizer.pth +1 -1
- a2c-AntBulletEnv-v0/policy.pth +1 -1
- a2c-AntBulletEnv-v0/system_info.txt +3 -3
- config.json +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
- vec_normalize.pkl +1 -1
.gitattributes
CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
+
replay.mp4 filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -16,7 +16,7 @@ model-index:
|
|
16 |
type: AntBulletEnv-v0
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
-
value:
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
16 |
type: AntBulletEnv-v0
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
+
value: 1163.43 +/- 305.61
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
a2c-AntBulletEnv-v0.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c98dca999a33346027d330cd14f55b40ed90be2e17a5981ceebd0b40f538d548
|
3 |
+
size 129281
|
a2c-AntBulletEnv-v0/data
CHANGED
@@ -4,20 +4,20 @@
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
11 |
-
"_build": "<function ActorCriticPolicy._build at
|
12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
-
"_abc_impl": "<_abc_data object at
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {
|
@@ -59,21 +59,21 @@
|
|
59 |
"_np_random": null
|
60 |
},
|
61 |
"n_envs": 4,
|
62 |
-
"num_timesteps":
|
63 |
"_total_timesteps": 2000000,
|
64 |
"_num_timesteps_at_start": 0,
|
65 |
"seed": null,
|
66 |
"action_noise": null,
|
67 |
-
"start_time":
|
68 |
"learning_rate": 0.00096,
|
69 |
"tensorboard_log": null,
|
70 |
"lr_schedule": {
|
71 |
":type:": "<class 'function'>",
|
72 |
-
":serialized:": "gAWVwwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+
|
73 |
},
|
74 |
"_last_obs": {
|
75 |
":type:": "<class 'numpy.ndarray'>",
|
76 |
-
":serialized:": "
|
77 |
},
|
78 |
"_last_episode_starts": {
|
79 |
":type:": "<class 'numpy.ndarray'>",
|
@@ -81,21 +81,21 @@
|
|
81 |
},
|
82 |
"_last_original_obs": {
|
83 |
":type:": "<class 'numpy.ndarray'>",
|
84 |
-
":serialized:": "
|
85 |
},
|
86 |
"_episode_num": 0,
|
87 |
"use_sde": true,
|
88 |
"sde_sample_freq": -1,
|
89 |
-
"_current_progress_remaining": 0.
|
90 |
"ep_info_buffer": {
|
91 |
":type:": "<class 'collections.deque'>",
|
92 |
-
":serialized:": "
|
93 |
},
|
94 |
"ep_success_buffer": {
|
95 |
":type:": "<class 'collections.deque'>",
|
96 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
97 |
},
|
98 |
-
"_n_updates":
|
99 |
"n_steps": 8,
|
100 |
"gamma": 0.99,
|
101 |
"gae_lambda": 0.9,
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7eff44164d30>",
|
8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7eff44164dc0>",
|
9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7eff44164e50>",
|
10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7eff44164ee0>",
|
11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7eff44164f70>",
|
12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7eff44168040>",
|
13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7eff441680d0>",
|
14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7eff44168160>",
|
15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7eff441681f0>",
|
16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7eff44168280>",
|
17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7eff44168310>",
|
18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7eff441683a0>",
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
+
"_abc_impl": "<_abc._abc_data object at 0x7eff4415f680>"
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {
|
59 |
"_np_random": null
|
60 |
},
|
61 |
"n_envs": 4,
|
62 |
+
"num_timesteps": 1074252,
|
63 |
"_total_timesteps": 2000000,
|
64 |
"_num_timesteps_at_start": 0,
|
65 |
"seed": null,
|
66 |
"action_noise": null,
|
67 |
+
"start_time": 1678764214657803571,
|
68 |
"learning_rate": 0.00096,
|
69 |
"tensorboard_log": null,
|
70 |
"lr_schedule": {
|
71 |
":type:": "<class 'function'>",
|
72 |
+
":serialized:": "gAWVwwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSC91c3IvbG9jYWwvbGliL3B5dGhvbjMuOS9kaXN0LXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZSMBGZ1bmOUS4JDAgABlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSC91c3IvbG9jYWwvbGliL3B5dGhvbjMuOS9kaXN0LXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgffZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/T3UQTVUdaYWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
73 |
},
|
74 |
"_last_obs": {
|
75 |
":type:": "<class 'numpy.ndarray'>",
|
76 |
+
":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAN+7MT+pw4s/sG7xPvJbBD7l6yW/ZuSwvo5GDb/RKPg+oPYvvwwYFsBdgE2/d5rav4ZCdb8WSum+K7mRvrwVeD8Be6G9cWAaQF+QXL+YT3C/NvVCPvpKar/wEzI/NVDbPki73L8uaYQ+umvsv/qxVD/9h94+Ktw0P9UtHj9u1wI/rInxPnCxIz9kXX69V3xCv+l7p7/Q5U0/342TP5N0Aj/5r5A/jfBkPwoHDT9iids8VHpuP3NcFb5s/Km+QjCVPq4AUj+pvEA+0o8BPv8rvb6pcxQ/LmmEPq6ZCj+RD5q/M0uYPtzZtT5NRCQ/5r0tPxzYMj1rpl4/TXEjvsjFQb9loiY+jOCOPwoYHb6PACpA/BYNv7nfmz8ekwk/prQtvywwkj96xzq94SHSvss1gj6v9G88gQmgP8QXkz6EJv+/SLvcvy5phD6umQo/kQ+av9D4Xr0bVQA/gekkP9lrm7+jqzK/qBtMPjfk7758W7C+QcYGwOoayz47T/g+3UbQvRdHmj+RTw89lYAOP7/0Oz2Xeo8/lS2BPgIUB7+ua26+SZOSP8EFJT/sNRg/zLA0PqlzFD8uaYQ+rpkKP/qxVD+UjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"
|
77 |
},
|
78 |
"_last_episode_starts": {
|
79 |
":type:": "<class 'numpy.ndarray'>",
|
81 |
},
|
82 |
"_last_original_obs": {
|
83 |
":type:": "<class 'numpy.ndarray'>",
|
84 |
+
":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAAAAACe8ys2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACA01C0vQAAAAB5ON+/AAAAAOIHsjwAAAAAXEP+PwAAAAAawIS6AAAAAEO39j8AAAAAiiMoOwAAAAAvc+G/AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA8Qs9tgAAgD8AAAAAAAAAAAAAAAAAAAAAAAAAgHqKJj0AAAAAG7TkvwAAAADrOBK9AAAAAMus5D8AAAAAEL64vQAAAABKIuU/AAAAABt1Sb0AAAAArIb/vwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAGkkpjYAAIA/AAAAAAAAAAAAAAAAAAAAAAAAAIBe6jO9AAAAAGez878AAAAAPZcIPgAAAACoJNw/AAAAAO2jjr0AAAAA8QjiPwAAAAAwO9O9AAAAAJkt6b8AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAIEJ22AACAPwAAAAAAAAAAAAAAAAAAAAAAAACASVGNPQAAAAB6T9m/AAAAALKwNb0AAAAABhnqPwAAAAB7wNy9AAAAAHNI2j8AAAAA5XGvvAAAAACIY/O/AAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"
|
85 |
},
|
86 |
"_episode_num": 0,
|
87 |
"use_sde": true,
|
88 |
"sde_sample_freq": -1,
|
89 |
+
"_current_progress_remaining": 0.46287999999999996,
|
90 |
"ep_info_buffer": {
|
91 |
":type:": "<class 'collections.deque'>",
|
92 |
+
":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQJo6XcafjCKMAWyUTegDjAF0lEdAnGp6iO/+KnV9lChoBkdAm27WI0qH5GgHTegDaAhHQJxzp1klNUR1fZQoaAZHQJvJjdvbXYloB03oA2gIR0CcfO9pRGc4dX2UKGgGR0CaB7W0Z3s5aAdN6ANoCEdAnILeK8+Ro3V9lChoBkdAmeM9Gd7OV2gHTegDaAhHQJyGtA9mpVF1fZQoaAZHQJpcg5WBBiVoB03oA2gIR0CcjKtCiRGMdX2UKGgGR0CZZmlA/s3RaAdN6ANoCEdAnJVOHFglW3V9lChoBkdAmzOjxoZhrmgHTegDaAhHQJybPr4WUKR1fZQoaAZHQJZUyMLncL1oB03oA2gIR0Ccn9rI5o4/dX2UKGgGR0CaO/k1uR9xaAdN6ANoCEdAnKgubqhUR3V9lChoBkdAmPD6akRBeGgHTegDaAhHQJy00E1VHWl1fZQoaAZHQJoG4gpz90loB03oA2gIR0Ccurdv863idX2UKGgGR0CXrctYSxqxaAdN6ANoCEdAnL6e/1xsEnV9lChoBkdAl3nm4uscQ2gHTegDaAhHQJzEYiW3Sa51fZQoaAZHQJloml54W1toB03oA2gIR0Cc0BM5OrQxdX2UKGgGR0CXUU8274BWaAdN6ANoCEdAnNk4UFjd6HV9lChoBkdAlzEPPTodMmgHTegDaAhHQJzfC/etSyd1fZQoaAZHQJowaj4593NoB03oA2gIR0Cc5+ttALRbdX2UKGgGR0CUhbYvnKW+aAdN6ANoCEdAnPOf/vOQhnV9lChoBkdAl5ki9ytFKGgHTegDaAhHQJz5jQ8fV7R1fZQoaAZHQJsxfr4WUKRoB03oA2gIR0Cc/YfUF0PpdX2UKGgGR0CaI8c580DVaAdN6ANoCEdAnQNXX2/SIHV9lChoBkdAnBgsfaHsTmgHTegDaAhHQJ0L3CZWq951fZQoaAZHQJkAOuV5a/1oB03oA2gIR0CdEeJPIn0DdX2UKGgGR0CZgd6u4gA7aAdN6ANoCEdAnRX0vwmVq3V9lChoBkdAnDXSIHkcTGgHTegDaAhHQJ0dJkoWpId1fZQoaAZHQJjIpO32EkBoB03oA2gIR0CdKi+m3vx6dX2UKGgGR0Ccyp7PY4ACaAdN6ANoCEdAnTEVcY64lXV9lChoBkdAnE5HUlRgqmgHTegDaAhHQJ01CQRwqAl1fZQoaAZHQJd4UWTHKfZoB03oA2gIR0CdOvFMIu5CdX2UKGgGR0CeBSHYYixFaAdN6ANoCEdAnUOUq+ajOHV9lChoBkdAmyfJuuRs/WgHTegDaAhHQJ1Jkq3Eycl1fZQoaAZHQJiQC8/UvwpoB03oA2gIR0CdTZ0Nz8xcdX2UKGgGR0CcZeu3c580aAdN6ANoCEdAnVOULH+6y3V9lChoBkdAneYuuFHrhWgHTegDaAhHQJ1fYs6JZW91fZQoaAZHQJRtovvjOs1oB03oA2gIR0CdaMP+GXXzdX2UKGgGR0CbzsJJGvwFaAdN6ANoCEdAnW0SnLq2SnV9lChoBkdAmYOvx2B8QmgHTegDaAhHQJ1y9t2s7uF1fZQoaAZHQJnN/DsMRYloB03oA2gIR0Cde5A08/2TdX2UKGgGR0CeTxA0sOG1aAdN6ANoCEdAnYFtB0IToXV9lChoBkdAkoHPvF3pwGgHTegDaAhHQJ2Fh13dKul1fZQoaAZHQJydb/YJ3PloB03oA2gIR0Cdiz814xDcdX2UKGgGR0CafKOPeYUnaAdN6ANoCEdAnZS9V/+bVnV9lChoBkdAm7YoFV1fV2gHTegDaAhHQJ2d2vdM0xd1fZQoaAZHQJuY+flIVdpoB03oA2gIR0Cdo/hwVCXydX2UKGgGR0CZ8GyAQQMAaAdN6ANoCEdAna02Y0EX+HV9lChoBkdAmOCkY0l7dGgHTegDaAhHQJ26jc9GI9F1fZQoaAZHQJwTC3nZCfJoB03oA2gIR0CdwLHLA57xdX2UKGgGR0CbnbkKu0TlaAdN6ANoCEdAncTDu0CzTnV9lChoBkdAmFHT4cm0FGgHTegDaAhHQJ3KtJJ5E+h1fZQoaAZHQJxEQPy08eVoB03oA2gIR0Cd04EaVD8cdX2UKGgGR0Ca/f+iJwbVaAdN6ANoCEdAndmObI91U3V9lChoBkdAmjMA93bEgmgHTegDaAhHQJ3dmosI3R51fZQoaAZHQJqI8CNjsldoB03oA2gIR0Cd5izErGzbdX2UKGgGR0CawYLMLWqcaAdN6ANoCEdAnfNY82aUinV9lChoBkdAnDv2cWj46GgHTegDaAhHQJ35ZNoJzDJ1fZQoaAZHQJoUjxVhkRVoB03oA2gIR0Cd/Wht+CsfdX2UKGgGR0CZpjU0elsQaAdN6ANoCEdAngNYIv8IiXV9lChoBkdAkruRQBPsRmgHTegDaAhHQJ4MDHktEoh1fZQoaAZHQJoVyZ1FH8VoB03oA2gIR0CeEi0G/vfCdX2UKGgGR0CYcGMHbAUMaAdN6ANoCEdAnhY3FPznR3V9lChoBkdAl6FloQFs6GgHTegDaAhHQJ4cdWZJCjV1fZQoaAZHQJj8+earmyRoB03oA2gIR0CeKVGMn7YTdX2UKGgGR0CZu8nm7rcCaAdN6ANoCEdAnjFvkBCD3HV9lChoBkdAmal8VtXPq2gHTegDaAhHQJ41UhJRO1x1fZQoaAZHQJuXSXsw+MZoB03oA2gIR0CeOyf8uSOjdX2UKGgGR0CaXLbZvkzXaAdN6ANoCEdAnkP5HEuQIXV9lChoBkdAmipC8J2MbWgHTegDaAhHQJ5JxPznRsx1fZQoaAZHQJpZ9WU8mrtoB03oA2gIR0CeTc02LpA2dX2UKGgGR0CZ0mGeMAFQaAdN6ANoCEdAnlO9Brvb5HV9lChoBkdAmhGYtxuKoGgHTegDaAhHQJ5eVLbpNbl1fZQoaAZHQJmOLbsWweNoB03oA2gIR0CeZ4jd56dEdX2UKGgGR0CZc80A93bFaAdN6ANoCEdAnm2at1ZDA3V9lChoBkdAmW2pPl+3IGgHTegDaAhHQJ51TUnXumd1fZQoaAZHQJiRB7XxvvVoB03oA2gIR0CegraRZEDydX2UKGgGR0CZPIv4/NaAaAdN6ANoCEdAnolYTPBzm3V9lChoBkdAmi6icPOIImgHTegDaAhHQJ6Nc3Jgb6x1fZQoaAZHQJZVtRm9QGhoB03oA2gIR0Cekz6Vt4zKdX2UKGgGR0CakeWTHKfWaAdN6ANoCEdAnp8EG3WnTHV9lChoBkdAmW56f8MuvmgHTegDaAhHQJ6oDUx20Rh1fZQoaAZHQJo3ES7GvOhoB03oA2gIR0CerKNh3JPqdX2UKGgGR0CaIFxt52QoaAdN6ANoCEdAnrKkRradtnV9lChoBkdAl1bsstkFwGgHTegDaAhHQJ67n70nPVx1fZQoaAZHQJLw7ALy+YdoB03oA2gIR0Cewarz5GjLdX2UKGgGR0CYA5P5YYBOaAdN6ANoCEdAnsWx6a9bo3V9lChoBkdAl9mgYgq3E2gHTegDaAhHQJ7Li3QUpNN1fZQoaAZHQJfDq3/givBoB03oA2gIR0Ce1cB1s+FDdX2UKGgGR0CZFeMcIZ62aAdN6ANoCEdAnt6MXN1QqXV9lChoBkdAl4lJM10knmgHTegDaAhHQJ7k0iqyWzF1fZQoaAZHQJXZqZDzAetoB03oA2gIR0Ce6y/jsD4hdX2UKGgGR0CY6OQiA2AHaAdN6ANoCEdAnvQQRXfZVXV9lChoBkdAmSQk65oXbmgHTegDaAhHQJ76MXBP9DR1fZQoaAZHQJhIA68xsVNoB03oA2gIR0Ce/k7PIGQkdX2UKGgGR0CYucH6uW8iaAdN6ANoCEdAnwQrtNSIg3V9lChoBkdAmY1LypaRp2gHTegDaAhHQJ8MwwWWQfZ1fZQoaAZHQJmwFqagElpoB03oA2gIR0CfFHX4TK1YdX2UKGgGR0CaJgvrnkksaAdN6ANoCEdAnxptytFKCnV9lChoBkdAmoe57TlT32gHTegDaAhHQJ8jUjTrmhd1fZQoaAZHQJsC9xT850doB03oA2gIR0CfLDTXJ5midX2UKGgGR0CYNJtPHktFaAdN6ANoCEdAnzJKCYkVvnVlLg=="
|
93 |
},
|
94 |
"ep_success_buffer": {
|
95 |
":type:": "<class 'collections.deque'>",
|
96 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
97 |
},
|
98 |
+
"_n_updates": 33570,
|
99 |
"n_steps": 8,
|
100 |
"gamma": 0.99,
|
101 |
"gae_lambda": 0.9,
|
a2c-AntBulletEnv-v0/policy.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 56190
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f4d77d0707e718a2c9c6ba6e83e4acce8aa7c7c7f227c896f0393e4d16dc23d
|
3 |
size 56190
|
a2c-AntBulletEnv-v0/policy.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 56958
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9815cafb6609f98d1a3230470928f8b481a1e6eddee7f358a9988533e163e14c
|
3 |
size 56958
|
a2c-AntBulletEnv-v0/system_info.txt
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
-
- OS: Linux-5.10.147+-x86_64-with-glibc2.
|
2 |
-
- Python: 3.
|
3 |
- Stable-Baselines3: 1.7.0
|
4 |
- PyTorch: 1.13.1+cu116
|
5 |
- GPU Enabled: True
|
6 |
-
- Numpy: 1.
|
7 |
- Gym: 0.21.0
|
1 |
+
- OS: Linux-5.10.147+-x86_64-with-glibc2.31 # 1 SMP Sat Dec 10 16:00:40 UTC 2022
|
2 |
+
- Python: 3.9.16
|
3 |
- Stable-Baselines3: 1.7.0
|
4 |
- PyTorch: 1.13.1+cu116
|
5 |
- GPU Enabled: True
|
6 |
+
- Numpy: 1.22.4
|
7 |
- Gym: 0.21.0
|
config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7fc38aa9a1f0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7fc38aa9a280>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7fc38aa9a310>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7fc38aa9a3a0>", "_build": "<function ActorCriticPolicy._build at 0x7fc38aa9a430>", "forward": "<function ActorCriticPolicy.forward at 0x7fc38aa9a4c0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7fc38aa9a550>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7fc38aa9a5e0>", "_predict": "<function ActorCriticPolicy._predict at 0x7fc38aa9a670>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7fc38aa9a700>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7fc38aa9a790>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7fc38aa9a820>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc_data object at 0x7fc38aa8dfc0>"}, "verbose": 1, "policy_kwargs": {":type:": "<class 'dict'>", ":serialized:": "gAWVowAAAAAAAAB9lCiMDGxvZ19zdGRfaW5pdJRK/v///4wKb3J0aG9faW5pdJSJjA9vcHRpbWl6ZXJfY2xhc3OUjBN0b3JjaC5vcHRpbS5ybXNwcm9wlIwHUk1TcHJvcJSTlIwQb3B0aW1pemVyX2t3YXJnc5R9lCiMBWFscGhhlEc/764UeuFHrowDZXBzlEc+5Pi1iONo8YwMd2VpZ2h0X2RlY2F5lEsAdXUu", "log_std_init": -2, "ortho_init": false, "optimizer_class": "<class 'torch.optim.rmsprop.RMSprop'>", "optimizer_kwargs": {"alpha": 0.99, "eps": 1e-05, "weight_decay": 0}}, "observation_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVZwIAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLHIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWcAAAAAAAAAAAAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/lGgKSxyFlIwBQ5R0lFKUjARoaWdolGgSKJZwAAAAAAAAAAAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH+UaApLHIWUaBV0lFKUjA1ib3VuZGVkX2JlbG93lGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLHIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaCFLHIWUaBV0lFKUjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "_shape": [28], "low": "[-inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf\n -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf]", "high": "[inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf\n inf inf inf inf inf inf inf inf inf inf]", "bounded_below": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "bounded_above": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "_np_random": null}, "action_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVnwEAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLCIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWIAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAv5RoCksIhZSMAUOUdJRSlIwEaGlnaJRoEiiWIAAAAAAAAAAAAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAP5RoCksIhZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolggAAAAAAAAAAQEBAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLCIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYIAAAAAAAAAAEBAQEBAQEBlGghSwiFlGgVdJRSlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "_shape": [8], "low": "[-1. -1. -1. -1. -1. -1. -1. -1.]", "high": "[1. 1. 1. 1. 1. 1. 1. 1.]", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_np_random": null}, "n_envs": 4, "num_timesteps": 2000000, "_total_timesteps": 2000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1675524532834137586, "learning_rate": 0.00096, "tensorboard_log": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVwwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSC91c3IvbG9jYWwvbGliL3B5dGhvbjMuOC9kaXN0LXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZSMBGZ1bmOUS4JDAgABlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSC91c3IvbG9jYWwvbGliL3B5dGhvbjMuOC9kaXN0LXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgffZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/T3UQTVUdaYWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="}, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAkper9v5Ng+ukrBPtUUjD+iQ2Y/dvrRvPwSyL4zyXM+HXaoveCLUz/Njw6/N+rqPkMh3T54YMo+iU1KPxGcjT8RwDg/O30cvyM7uT5cY5o/qByHP/rIz74YJLm+dNwbPx9kLT+Az6o+S/eYPpxE8j6zosy90TfHPytzd79niBU/2Nn1PxAXp75hqDk+FAZgPOJR179pXjQ9XxMKPgJp977STig+qKcOP5EvBT1e9ZC9H1fKvWmLw74L7nU+D98FvsI3RD9f1p295pvGvmSIHkAfZC0/gM+qPkv3mD5SQQfAZtO+v6DUKL6rkQ8/exonvyAKpL1p1AM+Oe+fvtCJSz+nBbs/jhy2PsvSa79Vpo+8E2RPvh/uCj/nOTs/WLAsOnfpwj62vwk/tX/1PsB2hzuSTJq+zxyDPtCpYb+28bK8H2QtP4DPqj5L95g+nETyPkKKwr8EGk++aHgQP9BBNb+k65a+jOTJPeZyv75zq0Q//iS+P/JlHL1eC3a/g2KjvKIef72ctm697/RMP5G3ET1fPvc+u1hJvUj78z7+5gs9rz2nvvDM+bxgDlu/qEGlvB9kLT+Az6o+S/eYPpxE8j6UjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYEAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwSFlIwBQ5R0lFKULg=="}, "_last_original_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAAAAAADioa1AACAPwAAAAAAAAAAAAAAAAAAAAAAAACAz1UGPAAAAAD9tvm/AAAAAPs1sz0AAAAASlTpPwAAAADhOVc8AAAAAKnD/j8AAAAAjXgGPgAAAADYvOS/AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAg9cDNgAAgD8AAAAAAAAAAAAAAAAAAAAAAAAAgAKQt70AAAAATBr7vwAAAAAUrjI9AAAAAAHc2T8AAAAAcoO2PAAAAADKY+E/AAAAAE0YmD0AAAAAMAvwvwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAADzZJzYAAIA/AAAAAAAAAAAAAAAAAAAAAAAAAIAc3Qq+AAAAAAjq7L8AAAAAATETvQAAAAAf1+s/AAAAABChpD0AAAAAewQAQAAAAAC4dJY9AAAAAGvS7r8AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAADfK4y1AACAPwAAAAAAAAAAAAAAAAAAAAAAAACAMULpPQAAAACuVdy/AAAAAOQRsTwAAAAAsCABQAAAAAAwX6M9AAAAAEyB8z8AAAAAzoMXPQAAAACXcPq/AAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_episode_num": 0, "use_sde": true, "sde_sample_freq": -1, "_current_progress_remaining": 0.0, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQJHL1zuF6AyMAWyUTegDjAF0lEdAqmkEB2fTTnV9lChoBkdAkEKjFZPl+2gHTegDaAhHQKpsa/M4cWF1fZQoaAZHQJA/XksBhhJoB03oA2gIR0CqbmvxH5JsdX2UKGgGR0CPZ2yTINmUaAdN6ANoCEdAqm9zjm0VrXV9lChoBkdAiIwna37UG2gHTegDaAhHQKp3kuQp4KR1fZQoaAZHQI9OngHeJpFoB03oA2gIR0CqedWE9MbndX2UKGgGR0CQeJ/6fra/aAdN6ANoCEdAqnsgODrZ8XV9lChoBkdAj4Kwmu1WsGgHTegDaAhHQKp70az/p+t1fZQoaAZHQJCPqjYZl4FoB03oA2gIR0Cqg8MJx//edX2UKGgGR0CQTnyLAHmjaAdN6ANoCEdAqobw+MZP23V9lChoBkdAiwawAEMb32gHTegDaAhHQKqI5K/20zF1fZQoaAZHQInC2LBKtgdoB03oA2gIR0CqigWUr08OdX2UKGgGR0CQs6n0Cih4aAdN6ANoCEdAqpNzDdgv13V9lChoBkdAkVxqG+K0lmgHTegDaAhHQKqVrNOdoWZ1fZQoaAZHQJC2XqyGBWhoB03oA2gIR0Cqlv/igkC4dX2UKGgGR0CRQRaYu01JaAdN6ANoCEdAqpevvWpZOnV9lChoBkdAkUoC+lCTlmgHTegDaAhHQKqftFYMfA91fZQoaAZHQJELC6nR9gFoB03oA2gIR0CqofBI4EOidX2UKGgGR0CRArdRR/EwaAdN6ANoCEdAqqOMo0ALiXV9lChoBkdAj6ywZXMhYGgHTegDaAhHQKqki/M4cWF1fZQoaAZHQJFWKig00nBoB03oA2gIR0Cqr1QsXizcdX2UKGgGR0CRjwk+X7cgaAdN6ANoCEdAqrGQNNJvpHV9lChoBkdAkMqZ2IO6NGgHTegDaAhHQKqy0LApKBd1fZQoaAZHQJDz45fdAPdoB03oA2gIR0Cqs4gDq4YrdX2UKGgGR0CRYmwHJLdvaAdN6ANoCEdAqrtT9deIEnV9lChoBkdAhrQZ6t1ZDGgHTegDaAhHQKq9kCA+Y+l1fZQoaAZHQIu1KaZx7zFoB03oA2gIR0CqvuYDcM3IdX2UKGgGR0CPwX0Lc9GJaAdN6ANoCEdAqr+Wd5IH1XV9lChoBkdAhm4Sk9ECvGgHTegDaAhHQKrKyAFPi1l1fZQoaAZHQIhtwNkOI69oB03oA2gIR0CqzWuJk5IZdX2UKGgGR0CNjydDIBBBaAdN6ANoCEdAqs6+kxh2GXV9lChoBkdAiBBimdiDumgHTegDaAhHQKrPb8tPHkt1fZQoaAZHQIod+HzpX6toB03oA2gIR0Cq13LTH80ldX2UKGgGR0CIqwWC2+fzaAdN6ANoCEdAqtnieZof0XV9lChoBkdAidAtQsPJ72gHTegDaAhHQKrbOGyon8d1fZQoaAZHQJHw6RcNYr9oB03oA2gIR0Cq2+Q176YWdX2UKGgGR0CQ3fB7NSqEaAdN6ANoCEdAquXGJaaCtnV9lChoBkdAiHxPTgEU02gHTegDaAhHQKrpP3/xUed1fZQoaAZHQJBf2ay8jA1oB03oA2gIR0Cq6tyLZSNwdX2UKGgGR0CKfBdi2DxtaAdN6ANoCEdAquuJyjpLVXV9lChoBkdAkUOVfiPyTmgHTegDaAhHQKrzS2sJY1Z1fZQoaAZHQI+6i0fHPu5oB03oA2gIR0Cq9YduHerNdX2UKGgGR0CTIQDlYEGJaAdN6ANoCEdAqvbUgfU4JnV9lChoBkdAj+2LdFfAsWgHTegDaAhHQKr3gf3evZB1fZQoaAZHQIwHSk9ECvJoB03oA2gIR0Cq/9i8vmHQdX2UKGgGR0CRUFMmWt2caAdN6ANoCEdAqwMutOmBOHV9lChoBkdAkF/Z/wy6+WgHTegDaAhHQKsFTKFqSHN1fZQoaAZHQIfG/ZmI0qJoB03oA2gIR0CrBmE0iyIIdX2UKGgGR0CLljst03fiaAdN6ANoCEdAqw756jWTYHV9lChoBkdAhxCxGc4HX2gHTegDaAhHQKsRUCEHt4R1fZQoaAZHQIu2d8Ti84BoB03oA2gIR0CrEpe8Gs3idX2UKGgGR0CLlQMyad+YaAdN6ANoCEdAqxNDdpItlXV9lChoBkdAkBZh2GIsRWgHTegDaAhHQKsa+hA4XGh1fZQoaAZHQJGR/qTr3TNoB03oA2gIR0CrHZblzU7TdX2UKGgGR0CNASkBS1mbaAdN6ANoCEdAqx9/Wz4UOHV9lChoBkdAjWdRDCxeLWgHTegDaAhHQKsgiA+Y+jd1fZQoaAZHQInaDL0SRKZoB03oA2gIR0CrKur4etCBdX2UKGgGR0CJFXYbsF+vaAdN6ANoCEdAqy09i2DxsnV9lChoBkdAgywoUBXCCWgHTegDaAhHQKsuisEq2Bt1fZQoaAZHQIsmkzoEB8xoB03oA2gIR0CrLzlBhQWOdX2UKGgGR0COuxK4hEBsaAdN6ANoCEdAqzbhYeT3ZnV9lChoBkdAkUUvgFX7tWgHTegDaAhHQKs5NllsguB1fZQoaAZHQIjMY+EAYHhoB03oA2gIR0CrOoRzRx95dX2UKGgGR0CKtDZf2K2saAdN6ANoCEdAqztBtJnQIHV9lChoBkdAiydMPJ7swGgHTegDaAhHQKtGeA/cFhZ1fZQoaAZHQJAsuiN83MpoB03oA2gIR0CrSMrE1l5GdX2UKGgGR0COxIQqZtvXaAdN6ANoCEdAq0oShL5AQnV9lChoBkdAjvRwgcLjP2gHTegDaAhHQKtKw78Nx2l1fZQoaAZHQJHxHOoo/iZoB03oA2gIR0CrUrTuOS4fdX2UKGgGR0CPfaqqfe1saAdN6ANoCEdAq1UJzT4L1HV9lChoBkdAkPtB4Uvf0mgHTegDaAhHQKtWWFBY3eh1fZQoaAZHQI+QkNH6MzdoB03oA2gIR0CrV05mh/RWdX2UKGgGR0CP2cbCrLhaaAdN6ANoCEdAq2I1+PRzBHV9lChoBkdAhnj2hh6SkmgHTegDaAhHQKtlLfEXLvF1fZQoaAZHQJEN5rpJPIpoB03oA2gIR0CrZoGmtQsPdX2UKGgGR0CJQcearmyPaAdN6ANoCEdAq2c3rt3OfXV9lChoBkdAivXRfWtlqmgHTegDaAhHQKtvHAO8TSN1fZQoaAZHQIaqg6wMYuVoB03oA2gIR0CrcXCZF5OadX2UKGgGR0CLCbGLk0aZaAdN6ANoCEdAq3LGEPDpDHV9lChoBkdAh65C2tuDSWgHTegDaAhHQKtze0Xxe9l1fZQoaAZHQIegx5qubI9oB03oA2gIR0CrfSj9GZuydX2UKGgGR0CGGzRqGlANaAdN6ANoCEdAq4DaCBf8dnV9lChoBkdAkbaN7SiM52gHTegDaAhHQKuCwH3UQTV1fZQoaAZHQIZ2bxCpm29oB03oA2gIR0Crg3W6bvw3dX2UKGgGR0CMUVo1UEPlaAdN6ANoCEdAq4t5esxO+XV9lChoBkdAiqUtzCDVY2gHTegDaAhHQKuNuWvbGm11fZQoaAZHQIy3XdqL0jFoB03oA2gIR0CrjweyiVSodX2UKGgGR0CMoprHlwLmaAdN6ANoCEdAq4+xXwLE1nV9lChoBkdAkRFZ+6RQrWgHTegDaAhHQKuYSuzyBkJ1fZQoaAZHQJEbY+t8uz1oB03oA2gIR0Crm6hTn7pFdX2UKGgGR0COt0OHWSU1aAdN6ANoCEdAq52nBFd9lXV9lChoBkdAhbYOtW+49WgHTegDaAhHQKuesqyWzGB1fZQoaAZHQI/g8RUWEbpoB03oA2gIR0Crp35SWJJodX2UKGgGR0CQSt0dilSCaAdN6ANoCEdAq6nFAAyVOnV9lChoBkdAkLVd38n/k2gHTegDaAhHQKurLM5fdAR1fZQoaAZHQI0xTg/C66JoB03oA2gIR0Crq98xKxs3dX2UKGgGR0CPkivXbuc+aAdN6ANoCEdAq7OvO6d1+3V9lChoBkdAh4Asoc7yQWgHTegDaAhHQKu2g5q/M4d1fZQoaAZHQIS1PHktEohoB03oA2gIR0CruGaBAfMfdX2UKGgGR0CRDZnctXgcaAdN6ANoCEdAq7lq3ocJdHVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 62500, "n_steps": 8, "gamma": 0.99, "gae_lambda": 0.9, "ent_coef": 0.0, "vf_coef": 0.4, "max_grad_norm": 0.5, "normalize_advantage": false, "system_info": {"OS": "Linux-5.10.147+-x86_64-with-glibc2.29 # 1 SMP Sat Dec 10 16:00:40 UTC 2022", "Python": "3.8.10", "Stable-Baselines3": "1.7.0", "PyTorch": "1.13.1+cu116", "GPU Enabled": "True", "Numpy": "1.21.6", "Gym": "0.21.0"}}
|
1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7eff44164d30>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7eff44164dc0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7eff44164e50>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7eff44164ee0>", "_build": "<function ActorCriticPolicy._build at 0x7eff44164f70>", "forward": "<function ActorCriticPolicy.forward at 0x7eff44168040>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7eff441680d0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7eff44168160>", "_predict": "<function ActorCriticPolicy._predict at 0x7eff441681f0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7eff44168280>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7eff44168310>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7eff441683a0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7eff4415f680>"}, "verbose": 1, "policy_kwargs": {":type:": "<class 'dict'>", ":serialized:": "gAWVowAAAAAAAAB9lCiMDGxvZ19zdGRfaW5pdJRK/v///4wKb3J0aG9faW5pdJSJjA9vcHRpbWl6ZXJfY2xhc3OUjBN0b3JjaC5vcHRpbS5ybXNwcm9wlIwHUk1TcHJvcJSTlIwQb3B0aW1pemVyX2t3YXJnc5R9lCiMBWFscGhhlEc/764UeuFHrowDZXBzlEc+5Pi1iONo8YwMd2VpZ2h0X2RlY2F5lEsAdXUu", "log_std_init": -2, "ortho_init": false, "optimizer_class": "<class 'torch.optim.rmsprop.RMSprop'>", "optimizer_kwargs": {"alpha": 0.99, "eps": 1e-05, "weight_decay": 0}}, "observation_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVZwIAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLHIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWcAAAAAAAAAAAAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/lGgKSxyFlIwBQ5R0lFKUjARoaWdolGgSKJZwAAAAAAAAAAAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH+UaApLHIWUaBV0lFKUjA1ib3VuZGVkX2JlbG93lGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLHIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaCFLHIWUaBV0lFKUjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "_shape": [28], "low": "[-inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf\n -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf]", "high": "[inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf\n inf inf inf inf inf inf inf inf inf inf]", "bounded_below": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "bounded_above": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "_np_random": null}, "action_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVnwEAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLCIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWIAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAv5RoCksIhZSMAUOUdJRSlIwEaGlnaJRoEiiWIAAAAAAAAAAAAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAP5RoCksIhZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolggAAAAAAAAAAQEBAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLCIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYIAAAAAAAAAAEBAQEBAQEBlGghSwiFlGgVdJRSlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "_shape": [8], "low": "[-1. -1. -1. -1. -1. -1. -1. -1.]", "high": "[1. 1. 1. 1. 1. 1. 1. 1.]", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_np_random": null}, "n_envs": 4, "num_timesteps": 1074252, "_total_timesteps": 2000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1678764214657803571, "learning_rate": 0.00096, "tensorboard_log": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVwwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSC91c3IvbG9jYWwvbGliL3B5dGhvbjMuOS9kaXN0LXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZSMBGZ1bmOUS4JDAgABlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSC91c3IvbG9jYWwvbGliL3B5dGhvbjMuOS9kaXN0LXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZR1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgffZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/T3UQTVUdaYWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="}, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAN+7MT+pw4s/sG7xPvJbBD7l6yW/ZuSwvo5GDb/RKPg+oPYvvwwYFsBdgE2/d5rav4ZCdb8WSum+K7mRvrwVeD8Be6G9cWAaQF+QXL+YT3C/NvVCPvpKar/wEzI/NVDbPki73L8uaYQ+umvsv/qxVD/9h94+Ktw0P9UtHj9u1wI/rInxPnCxIz9kXX69V3xCv+l7p7/Q5U0/342TP5N0Aj/5r5A/jfBkPwoHDT9iids8VHpuP3NcFb5s/Km+QjCVPq4AUj+pvEA+0o8BPv8rvb6pcxQ/LmmEPq6ZCj+RD5q/M0uYPtzZtT5NRCQ/5r0tPxzYMj1rpl4/TXEjvsjFQb9loiY+jOCOPwoYHb6PACpA/BYNv7nfmz8ekwk/prQtvywwkj96xzq94SHSvss1gj6v9G88gQmgP8QXkz6EJv+/SLvcvy5phD6umQo/kQ+av9D4Xr0bVQA/gekkP9lrm7+jqzK/qBtMPjfk7758W7C+QcYGwOoayz47T/g+3UbQvRdHmj+RTw89lYAOP7/0Oz2Xeo8/lS2BPgIUB7+ua26+SZOSP8EFJT/sNRg/zLA0PqlzFD8uaYQ+rpkKP/qxVD+UjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYEAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwSFlIwBQ5R0lFKULg=="}, "_last_original_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAAAAACe8ys2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACA01C0vQAAAAB5ON+/AAAAAOIHsjwAAAAAXEP+PwAAAAAawIS6AAAAAEO39j8AAAAAiiMoOwAAAAAvc+G/AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA8Qs9tgAAgD8AAAAAAAAAAAAAAAAAAAAAAAAAgHqKJj0AAAAAG7TkvwAAAADrOBK9AAAAAMus5D8AAAAAEL64vQAAAABKIuU/AAAAABt1Sb0AAAAArIb/vwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAGkkpjYAAIA/AAAAAAAAAAAAAAAAAAAAAAAAAIBe6jO9AAAAAGez878AAAAAPZcIPgAAAACoJNw/AAAAAO2jjr0AAAAA8QjiPwAAAAAwO9O9AAAAAJkt6b8AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAIEJ22AACAPwAAAAAAAAAAAAAAAAAAAAAAAACASVGNPQAAAAB6T9m/AAAAALKwNb0AAAAABhnqPwAAAAB7wNy9AAAAAHNI2j8AAAAA5XGvvAAAAACIY/O/AAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_episode_num": 0, "use_sde": true, "sde_sample_freq": -1, "_current_progress_remaining": 0.46287999999999996, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQJo6XcafjCKMAWyUTegDjAF0lEdAnGp6iO/+KnV9lChoBkdAm27WI0qH5GgHTegDaAhHQJxzp1klNUR1fZQoaAZHQJvJjdvbXYloB03oA2gIR0CcfO9pRGc4dX2UKGgGR0CaB7W0Z3s5aAdN6ANoCEdAnILeK8+Ro3V9lChoBkdAmeM9Gd7OV2gHTegDaAhHQJyGtA9mpVF1fZQoaAZHQJpcg5WBBiVoB03oA2gIR0CcjKtCiRGMdX2UKGgGR0CZZmlA/s3RaAdN6ANoCEdAnJVOHFglW3V9lChoBkdAmzOjxoZhrmgHTegDaAhHQJybPr4WUKR1fZQoaAZHQJZUyMLncL1oB03oA2gIR0Ccn9rI5o4/dX2UKGgGR0CaO/k1uR9xaAdN6ANoCEdAnKgubqhUR3V9lChoBkdAmPD6akRBeGgHTegDaAhHQJy00E1VHWl1fZQoaAZHQJoG4gpz90loB03oA2gIR0Ccurdv863idX2UKGgGR0CXrctYSxqxaAdN6ANoCEdAnL6e/1xsEnV9lChoBkdAl3nm4uscQ2gHTegDaAhHQJzEYiW3Sa51fZQoaAZHQJloml54W1toB03oA2gIR0Cc0BM5OrQxdX2UKGgGR0CXUU8274BWaAdN6ANoCEdAnNk4UFjd6HV9lChoBkdAlzEPPTodMmgHTegDaAhHQJzfC/etSyd1fZQoaAZHQJowaj4593NoB03oA2gIR0Cc5+ttALRbdX2UKGgGR0CUhbYvnKW+aAdN6ANoCEdAnPOf/vOQhnV9lChoBkdAl5ki9ytFKGgHTegDaAhHQJz5jQ8fV7R1fZQoaAZHQJsxfr4WUKRoB03oA2gIR0Cc/YfUF0PpdX2UKGgGR0CaI8c580DVaAdN6ANoCEdAnQNXX2/SIHV9lChoBkdAnBgsfaHsTmgHTegDaAhHQJ0L3CZWq951fZQoaAZHQJkAOuV5a/1oB03oA2gIR0CdEeJPIn0DdX2UKGgGR0CZgd6u4gA7aAdN6ANoCEdAnRX0vwmVq3V9lChoBkdAnDXSIHkcTGgHTegDaAhHQJ0dJkoWpId1fZQoaAZHQJjIpO32EkBoB03oA2gIR0CdKi+m3vx6dX2UKGgGR0Ccyp7PY4ACaAdN6ANoCEdAnTEVcY64lXV9lChoBkdAnE5HUlRgqmgHTegDaAhHQJ01CQRwqAl1fZQoaAZHQJd4UWTHKfZoB03oA2gIR0CdOvFMIu5CdX2UKGgGR0CeBSHYYixFaAdN6ANoCEdAnUOUq+ajOHV9lChoBkdAmyfJuuRs/WgHTegDaAhHQJ1Jkq3Eycl1fZQoaAZHQJiQC8/UvwpoB03oA2gIR0CdTZ0Nz8xcdX2UKGgGR0CcZeu3c580aAdN6ANoCEdAnVOULH+6y3V9lChoBkdAneYuuFHrhWgHTegDaAhHQJ1fYs6JZW91fZQoaAZHQJRtovvjOs1oB03oA2gIR0CdaMP+GXXzdX2UKGgGR0CbzsJJGvwFaAdN6ANoCEdAnW0SnLq2SnV9lChoBkdAmYOvx2B8QmgHTegDaAhHQJ1y9t2s7uF1fZQoaAZHQJnN/DsMRYloB03oA2gIR0Cde5A08/2TdX2UKGgGR0CeTxA0sOG1aAdN6ANoCEdAnYFtB0IToXV9lChoBkdAkoHPvF3pwGgHTegDaAhHQJ2Fh13dKul1fZQoaAZHQJydb/YJ3PloB03oA2gIR0Cdiz814xDcdX2UKGgGR0CafKOPeYUnaAdN6ANoCEdAnZS9V/+bVnV9lChoBkdAm7YoFV1fV2gHTegDaAhHQJ2d2vdM0xd1fZQoaAZHQJuY+flIVdpoB03oA2gIR0Cdo/hwVCXydX2UKGgGR0CZ8GyAQQMAaAdN6ANoCEdAna02Y0EX+HV9lChoBkdAmOCkY0l7dGgHTegDaAhHQJ26jc9GI9F1fZQoaAZHQJwTC3nZCfJoB03oA2gIR0CdwLHLA57xdX2UKGgGR0CbnbkKu0TlaAdN6ANoCEdAncTDu0CzTnV9lChoBkdAmFHT4cm0FGgHTegDaAhHQJ3KtJJ5E+h1fZQoaAZHQJxEQPy08eVoB03oA2gIR0Cd04EaVD8cdX2UKGgGR0Ca/f+iJwbVaAdN6ANoCEdAndmObI91U3V9lChoBkdAmjMA93bEgmgHTegDaAhHQJ3dmosI3R51fZQoaAZHQJqI8CNjsldoB03oA2gIR0Cd5izErGzbdX2UKGgGR0CawYLMLWqcaAdN6ANoCEdAnfNY82aUinV9lChoBkdAnDv2cWj46GgHTegDaAhHQJ35ZNoJzDJ1fZQoaAZHQJoUjxVhkRVoB03oA2gIR0Cd/Wht+CsfdX2UKGgGR0CZpjU0elsQaAdN6ANoCEdAngNYIv8IiXV9lChoBkdAkruRQBPsRmgHTegDaAhHQJ4MDHktEoh1fZQoaAZHQJoVyZ1FH8VoB03oA2gIR0CeEi0G/vfCdX2UKGgGR0CYcGMHbAUMaAdN6ANoCEdAnhY3FPznR3V9lChoBkdAl6FloQFs6GgHTegDaAhHQJ4cdWZJCjV1fZQoaAZHQJj8+earmyRoB03oA2gIR0CeKVGMn7YTdX2UKGgGR0CZu8nm7rcCaAdN6ANoCEdAnjFvkBCD3HV9lChoBkdAmal8VtXPq2gHTegDaAhHQJ41UhJRO1x1fZQoaAZHQJuXSXsw+MZoB03oA2gIR0CeOyf8uSOjdX2UKGgGR0CaXLbZvkzXaAdN6ANoCEdAnkP5HEuQIXV9lChoBkdAmipC8J2MbWgHTegDaAhHQJ5JxPznRsx1fZQoaAZHQJpZ9WU8mrtoB03oA2gIR0CeTc02LpA2dX2UKGgGR0CZ0mGeMAFQaAdN6ANoCEdAnlO9Brvb5HV9lChoBkdAmhGYtxuKoGgHTegDaAhHQJ5eVLbpNbl1fZQoaAZHQJmOLbsWweNoB03oA2gIR0CeZ4jd56dEdX2UKGgGR0CZc80A93bFaAdN6ANoCEdAnm2at1ZDA3V9lChoBkdAmW2pPl+3IGgHTegDaAhHQJ51TUnXumd1fZQoaAZHQJiRB7XxvvVoB03oA2gIR0CegraRZEDydX2UKGgGR0CZPIv4/NaAaAdN6ANoCEdAnolYTPBzm3V9lChoBkdAmi6icPOIImgHTegDaAhHQJ6Nc3Jgb6x1fZQoaAZHQJZVtRm9QGhoB03oA2gIR0Cekz6Vt4zKdX2UKGgGR0CakeWTHKfWaAdN6ANoCEdAnp8EG3WnTHV9lChoBkdAmW56f8MuvmgHTegDaAhHQJ6oDUx20Rh1fZQoaAZHQJo3ES7GvOhoB03oA2gIR0CerKNh3JPqdX2UKGgGR0CaIFxt52QoaAdN6ANoCEdAnrKkRradtnV9lChoBkdAl1bsstkFwGgHTegDaAhHQJ67n70nPVx1fZQoaAZHQJLw7ALy+YdoB03oA2gIR0Cewarz5GjLdX2UKGgGR0CYA5P5YYBOaAdN6ANoCEdAnsWx6a9bo3V9lChoBkdAl9mgYgq3E2gHTegDaAhHQJ7Li3QUpNN1fZQoaAZHQJfDq3/givBoB03oA2gIR0Ce1cB1s+FDdX2UKGgGR0CZFeMcIZ62aAdN6ANoCEdAnt6MXN1QqXV9lChoBkdAl4lJM10knmgHTegDaAhHQJ7k0iqyWzF1fZQoaAZHQJXZqZDzAetoB03oA2gIR0Ce6y/jsD4hdX2UKGgGR0CY6OQiA2AHaAdN6ANoCEdAnvQQRXfZVXV9lChoBkdAmSQk65oXbmgHTegDaAhHQJ76MXBP9DR1fZQoaAZHQJhIA68xsVNoB03oA2gIR0Ce/k7PIGQkdX2UKGgGR0CYucH6uW8iaAdN6ANoCEdAnwQrtNSIg3V9lChoBkdAmY1LypaRp2gHTegDaAhHQJ8MwwWWQfZ1fZQoaAZHQJmwFqagElpoB03oA2gIR0CfFHX4TK1YdX2UKGgGR0CaJgvrnkksaAdN6ANoCEdAnxptytFKCnV9lChoBkdAmoe57TlT32gHTegDaAhHQJ8jUjTrmhd1fZQoaAZHQJsC9xT850doB03oA2gIR0CfLDTXJ5midX2UKGgGR0CYNJtPHktFaAdN6ANoCEdAnzJKCYkVvnVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 33570, "n_steps": 8, "gamma": 0.99, "gae_lambda": 0.9, "ent_coef": 0.0, "vf_coef": 0.4, "max_grad_norm": 0.5, "normalize_advantage": false, "system_info": {"OS": "Linux-5.10.147+-x86_64-with-glibc2.31 # 1 SMP Sat Dec 10 16:00:40 UTC 2022", "Python": "3.9.16", "Stable-Baselines3": "1.7.0", "PyTorch": "1.13.1+cu116", "GPU Enabled": "True", "Numpy": "1.22.4", "Gym": "0.21.0"}}
|
replay.mp4
CHANGED
Binary files a/replay.mp4 and b/replay.mp4 differ
|
results.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"mean_reward":
|
1 |
+
{"mean_reward": 1163.4319134610473, "std_reward": 305.6091817931826, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-03-14T04:01:24.722443"}
|
vec_normalize.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2136
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29796b5671dc1ad4c133679cd0b6470d912f2b59145bbf9c18545b7e94a87be2
|
3 |
size 2136
|