Initial commit
Browse files- README.md +1 -1
- a2c-AntBulletEnv-v0.zip +2 -2
- a2c-AntBulletEnv-v0/data +20 -20
- a2c-AntBulletEnv-v0/policy.optimizer.pth +1 -1
- a2c-AntBulletEnv-v0/policy.pth +1 -1
- config.json +1 -1
- replay.mp4 +2 -2
- results.json +1 -1
README.md
CHANGED
@@ -16,7 +16,7 @@ model-index:
|
|
16 |
type: AntBulletEnv-v0
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
-
value:
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
|
|
16 |
type: AntBulletEnv-v0
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
+
value: 859.12 +/- 326.92
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
a2c-AntBulletEnv-v0.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5caba383005a24026168241f57568347a05db7c9a37bd16bcb92a48e30895b9f
|
3 |
+
size 129275
|
a2c-AntBulletEnv-v0/data
CHANGED
@@ -4,20 +4,20 @@
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
11 |
-
"_build": "<function ActorCriticPolicy._build at
|
12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
-
"_abc_impl": "<_abc._abc_data object at
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {
|
@@ -59,12 +59,12 @@
|
|
59 |
"_np_random": null
|
60 |
},
|
61 |
"n_envs": 4,
|
62 |
-
"num_timesteps":
|
63 |
-
"_total_timesteps":
|
64 |
"_num_timesteps_at_start": 0,
|
65 |
"seed": null,
|
66 |
"action_noise": null,
|
67 |
-
"start_time":
|
68 |
"learning_rate": 0.00096,
|
69 |
"tensorboard_log": null,
|
70 |
"lr_schedule": {
|
@@ -73,7 +73,7 @@
|
|
73 |
},
|
74 |
"_last_obs": {
|
75 |
":type:": "<class 'numpy.ndarray'>",
|
76 |
-
":serialized:": "
|
77 |
},
|
78 |
"_last_episode_starts": {
|
79 |
":type:": "<class 'numpy.ndarray'>",
|
@@ -81,7 +81,7 @@
|
|
81 |
},
|
82 |
"_last_original_obs": {
|
83 |
":type:": "<class 'numpy.ndarray'>",
|
84 |
-
":serialized:": "
|
85 |
},
|
86 |
"_episode_num": 0,
|
87 |
"use_sde": true,
|
@@ -89,13 +89,13 @@
|
|
89 |
"_current_progress_remaining": 0.0,
|
90 |
"ep_info_buffer": {
|
91 |
":type:": "<class 'collections.deque'>",
|
92 |
-
":serialized:": "
|
93 |
},
|
94 |
"ep_success_buffer": {
|
95 |
":type:": "<class 'collections.deque'>",
|
96 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
97 |
},
|
98 |
-
"_n_updates":
|
99 |
"n_steps": 8,
|
100 |
"gamma": 0.99,
|
101 |
"gae_lambda": 0.9,
|
|
|
4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
5 |
"__module__": "stable_baselines3.common.policies",
|
6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7f9efa3f1ea0>",
|
8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f9efa3f1f30>",
|
9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f9efa3f1fc0>",
|
10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f9efa3f2050>",
|
11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7f9efa3f20e0>",
|
12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7f9efa3f2170>",
|
13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7f9efa3f2200>",
|
14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f9efa3f2290>",
|
15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7f9efa3f2320>",
|
16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f9efa3f23b0>",
|
17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f9efa3f2440>",
|
18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7f9efa3f24d0>",
|
19 |
"__abstractmethods__": "frozenset()",
|
20 |
+
"_abc_impl": "<_abc._abc_data object at 0x7f9efa3f8700>"
|
21 |
},
|
22 |
"verbose": 1,
|
23 |
"policy_kwargs": {
|
|
|
59 |
"_np_random": null
|
60 |
},
|
61 |
"n_envs": 4,
|
62 |
+
"num_timesteps": 2000000,
|
63 |
+
"_total_timesteps": 2000000,
|
64 |
"_num_timesteps_at_start": 0,
|
65 |
"seed": null,
|
66 |
"action_noise": null,
|
67 |
+
"start_time": 1676881276006055953,
|
68 |
"learning_rate": 0.00096,
|
69 |
"tensorboard_log": null,
|
70 |
"lr_schedule": {
|
|
|
73 |
},
|
74 |
"_last_obs": {
|
75 |
":type:": "<class 'numpy.ndarray'>",
|
76 |
+
":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAADoLK0Bn8gFAVV5AwMobp77OPhHAVT8MQFRfHEDT9ou/PkSRwNrTbT/cY8a+kbLqPp4a/T/Ziu2+Q6QHwJCoRz/xrRk/46w6vytRsz4bsIK/IC60Pt0tdsDkWARA/eurP6pEqr88xhY/ZdjmPoW9VD/LeDk/UW9pvyZ5yz5rAE8/ldoxv6mzRr9o+ok79hwpvz4+gj6uh7k+B/+8vjmgTMAi4xu/A+CmP81Cx70Kol4+odvzPiZ+7D+AF4E/s/KwvR28FL/PGrk9SWI4PzZ4TkCqRKq/1lTZv6jyDcA1B5q/O+LEP9HQib7RAgk/ZkkUP4kUCsAK7cs/HK4owIQDmb6yR4HAmTOLPl+AdT+rZ9M+yMOtv6FGVj187zk/gsLqvE1ECcBs9a2/vhoQwJS8gz+CWBS9cncCwJRSCT50nvO+BXNAPzzGFj9l2OY+NQeavyTChL/SYfi/JiELvzIzSr/n+RC/hN3uvsN4y7204O29/UYDP6a0br8oh8m++3p7v7ffxz4GrCS+2wMtPwgXZj4ToFs/8Tp8Pl5WFT/bLIq/FgRRP+uPtrv6ZTu+zT1GPgVzQD88xhY/ZdjmPjUHmr+UjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"
|
77 |
},
|
78 |
"_last_episode_starts": {
|
79 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
81 |
},
|
82 |
"_last_original_obs": {
|
83 |
":type:": "<class 'numpy.ndarray'>",
|
84 |
+
":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAAAAABMklW2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACAMKWyvQAAAAA5sOq/AAAAAHqagj0AAAAAxozvPwAAAAAJfgu+AAAAAATh4j8AAAAAm0YRPgAAAABDE/S/AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAX9zutgAAgD8AAAAAAAAAAAAAAAAAAAAAAAAAgBPUB7wAAAAAvxHrvwAAAAC+swC+AAAAACItAEAAAAAAjOZzvQAAAABtG+w/AAAAAFhUBD4AAAAA7+TkvwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAF3uNzYAAIA/AAAAAAAAAAAAAAAAAAAAAAAAAIAVhW49AAAAAH2q3r8AAAAAkpvuPQAAAAAsxfI/AAAAAF7bkjwAAAAAxID9PwAAAACHLBW9AAAAAFdV8r8AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAABQvsS2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACA+C6kPQAAAAD0Wfy/AAAAAI2Sw7sAAAAAVTP2PwAAAABFVCK9AAAAANP68T8AAAAABFb9PQAAAACJE/e/AAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"
|
85 |
},
|
86 |
"_episode_num": 0,
|
87 |
"use_sde": true,
|
|
|
89 |
"_current_progress_remaining": 0.0,
|
90 |
"ep_info_buffer": {
|
91 |
":type:": "<class 'collections.deque'>",
|
92 |
+
":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQIfbrqfOD8OMAWyUTegDjAF0lEdAoLlIDgZTAHV9lChoBkdAh98CYkVvdmgHTegDaAhHQKC6lUutfXx1fZQoaAZHQIsFS2hIvrZoB03oA2gIR0Cgu/GMOwxGdX2UKGgGR0CL0qVC5VfeaAdN6ANoCEdAoLwmZiNKiHV9lChoBkdAkvVLmZE2HmgHTegDaAhHQKDCKr6LwWp1fZQoaAZHQJOCgWP91lpoB03oA2gIR0Cgw3TrVvuPdX2UKGgGR0CNwraZhKDkaAdN6ANoCEdAoMTNkauOj3V9lChoBkdAlXBeEh7mdWgHTegDaAhHQKDFAjL0SRN1fZQoaAZHQJCLS6I3zc1oB03oA2gIR0CgyxmwJPZadX2UKGgGR0CKUy+mFajfaAdN6ANoCEdAoMxoaNuLrHV9lChoBkdAiAfRtxdY4mgHTegDaAhHQKDNvvIfbK11fZQoaAZHQIiqNb7j1f5oB03oA2gIR0CgzfM9KVY7dX2UKGgGR0CKMltzjm0WaAdN6ANoCEdAoNQdzbN8mnV9lChoBkdAiGUXE61b7mgHTegDaAhHQKDVdZowmE51fZQoaAZHQIR99FnZkCpoB03oA2gIR0Cg1tYXfqHHdX2UKGgGR0CPpIdNFjNIaAdN6ANoCEdAoNcLN6gM+nV9lChoBkdAh6l3OGCZnmgHTegDaAhHQKDdNiqhlDp1fZQoaAZHQIbMLCcf/3poB03oA2gIR0Cg3oTollbvdX2UKGgGR0CKityGzru6aAdN6ANoCEdAoN/eR7qptXV9lChoBkdAiuQXHaN+9mgHTegDaAhHQKDgE3uuzQh1fZQoaAZHQIoG4RTS9dxoB03oA2gIR0Cg5i4DTz/ZdX2UKGgGR0CLX2u/UONHaAdN6ANoCEdAoOd5/G2kSHV9lChoBkdAhxRiCz1K5GgHTegDaAhHQKDo0hQm/nJ1fZQoaAZHQIrVI9V3ljpoB03oA2gIR0Cg6QcZLqUvdX2UKGgGR0CKq9Ys/Y8MaAdN6ANoCEdAoO8cDB/I83V9lChoBkdAjZzgUUO/cmgHTegDaAhHQKDwZLeQ+2V1fZQoaAZHQJJ4/a24NI9oB03oA2gIR0Cg8b1GTcIrdX2UKGgGR0CM/+mtyPuHaAdN6ANoCEdAoPHxkI5YHXV9lChoBkdAjXyYMnZ00WgHTegDaAhHQKD4CpiqhlF1fZQoaAZHQIkhS9kBjnVoB03oA2gIR0Cg+VOoo/iYdX2UKGgGR0CKGxg4OtnxaAdN6ANoCEdAoPqvrY5DJHV9lChoBkdAiRPGsmv4d2gHTegDaAhHQKD65MV1wHZ1fZQoaAZHQIk2DKvFFUhoB03oA2gIR0ChAP1SwW30dX2UKGgGR0CQqFIMSbpeaAdN6ANoCEdAoQJHpdKNAHV9lChoBkdAh40Zk9U0emgHTegDaAhHQKEDnnFHavl1fZQoaAZHQI+v26TW5H5oB03oA2gIR0ChA9Mmv4dqdX2UKGgGR0CF12tmthd/aAdN6ANoCEdAoQnp15jYqXV9lChoBkdAkOmCzcAR02gHTegDaAhHQKELM/BWPtF1fZQoaAZHQIkNRiy6cy5oB03oA2gIR0ChDI3EyckMdX2UKGgGR0CQBtk1Mue0aAdN6ANoCEdAoQzCBRQ793V9lChoBkdAhkxcVYZEUmgHTegDaAhHQKES3kuHvc91fZQoaAZHQIYodqUNayNoB03oA2gIR0ChFCaX8fmtdX2UKGgGR0CTj4HJcPe6aAdN6ANoCEdAoRV9r2xptnV9lChoBkdAjixbLt/nXGgHTegDaAhHQKEVsSs8xKx1fZQoaAZHQIw617rs0HhoB03oA2gIR0ChG8eBH09RdX2UKGgGR0CLuXLPldTpaAdN6ANoCEdAoR0UeMhounV9lChoBkdAhcYrHEMspWgHTegDaAhHQKEecFK02Lp1fZQoaAZHQIl21/OMVDdoB03oA2gIR0ChHqTvJA+qdX2UKGgGR0CTGTsRxtHhaAdN6ANoCEdAoSS3BguyvHV9lChoBkdAinqWrfcesGgHTegDaAhHQKEmAq3mV7h1fZQoaAZHQJMFKUr08NhoB03oA2gIR0ChJ1l9jPOZdX2UKGgGR0CLY/a9sabXaAdN6ANoCEdAoSeOLpA2RHV9lChoBkdAk/ZG43FUAGgHTegDaAhHQKEtjqqwQlN1fZQoaAZHQJKYPEuQIUtoB03oA2gIR0ChLtpZOi35dX2UKGgGR0CQW7qYJE6UaAdN6ANoCEdAoTAxDNQj2XV9lChoBkdAk33pqREF4mgHTegDaAhHQKEwZLrX18N1fZQoaAZHQJHhiU7jkuJoB03oA2gIR0ChNmdNN8E3dX2UKGgGR0CRVX1L8JlbaAdN6ANoCEdAoTev6TGHYnV9lChoBkdAklCS+UQkHGgHTegDaAhHQKE5BrAP/aR1fZQoaAZHQJHnxhvze41oB03oA2gIR0ChOTtQKrq/dX2UKGgGR0CThMMbWEsbaAdN6ANoCEdAoT9Fa8pTdnV9lChoBkdAkP6HW8RL9WgHTegDaAhHQKFAj8rqdH51fZQoaAZHQI6Adg4OtnxoB03oA2gIR0ChQeSH2ys0dX2UKGgGR0CVMUDEFW4maAdN6ANoCEdAoUIY7xNIsnV9lChoBkdAkzjmn0kGA2gHTegDaAhHQKFIFqrzXjF1fZQoaAZHQJTvoRChN/RoB03oA2gIR0ChSV0jLSuydX2UKGgGR0CPWpyWAwwkaAdN6ANoCEdAoUq1qL0jDHV9lChoBkdAlLlv/NqxkmgHTegDaAhHQKFK6eQMhHN1fZQoaAZHQJR1CGpMpPRoB03oA2gIR0ChUPdznzQNdX2UKGgGR0CR8LvCdjG2aAdN6ANoCEdAoVJAgxJumHV9lChoBkdAlHzGJzkp7WgHTegDaAhHQKFTlszEaVF1fZQoaAZHQJQu73ztkWhoB03oA2gIR0ChU8sefZmJdX2UKGgGR0CTc3phF3INaAdN6ANoCEdAoVneMbWEsnV9lChoBkdAlYFFHFxXGWgHTegDaAhHQKFbKCz1K5F1fZQoaAZHQJEcmy2QXANoB03oA2gIR0ChXH9weeWfdX2UKGgGR0CMs9M5fdAPaAdN6ANoCEdAoVyzyDqW1XV9lChoBkdAkcuoHPeHi2gHTegDaAhHQKFivCvX9R91fZQoaAZHQInX+uRs/INoB03oA2gIR0ChZASaEzwddX2UKGgGR0CUHVc/+sHTaAdN6ANoCEdAoWVcMiKR+3V9lChoBkdAkNCZMQEpzGgHTegDaAhHQKFlkHKOktV1fZQoaAZHQI8zDjYI0IloB03oA2gIR0Cha5hhYvFndX2UKGgGR0CTI5i704BFaAdN6ANoCEdAoWzehXbM5nV9lChoBkdAkke7R8c+7mgHTegDaAhHQKFuMmsNlRR1fZQoaAZHQJSvLqoqCpZoB03oA2gIR0ChbmY0VJtjdX2UKGgGR0CWEztj0+TvaAdN6ANoCEdAoXRj6ciGFnV9lChoBkdAlHwgzP8htGgHTegDaAhHQKF1p1loUSJ1fZQoaAZHQJODk8nuy/toB03oA2gIR0ChdvuxrzoVdX2UKGgGR0CUL6LRa5f/aAdN6ANoCEdAoXcv3Dej23V9lChoBkdAit3q6nR9gGgHTegDaAhHQKF9Pe7+T/11fZQoaAZHQJSQTKB/ZuhoB03oA2gIR0ChfoSOaOPvdX2UKGgGR0CR4gU+LWI5aAdN6ANoCEdAoX/Zjvuw5nV9lChoBkdAktoDF+/gzmgHTegDaAhHQKGADORkmQd1fZQoaAZHQJbUj9ZRsM1oB03oA2gIR0Chhg0oa1kUdX2UKGgGR0CWbi44p+c6aAdN6ANoCEdAoYdT3PAwf3V9lChoBkdAjL4/IsAeaWgHTegDaAhHQKGIqKjzqbB1fZQoaAZHQJKcaxGDtgNoB03oA2gIR0ChiNyB06o3dX2UKGgGR0CWHQmbb1yvaAdN6ANoCEdAoY7ULv1DjXV9lChoBkdAk8OYK+i8F2gHTegDaAhHQKGQF0se4kN1fZQoaAZHQJWHovIwM6RoB03oA2gIR0ChkWkzfrKOdX2UKGgGR0CSPfom5UcXaAdN6ANoCEdAoZGcewLVnXVlLg=="
|
93 |
},
|
94 |
"ep_success_buffer": {
|
95 |
":type:": "<class 'collections.deque'>",
|
96 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
97 |
},
|
98 |
+
"_n_updates": 62500,
|
99 |
"n_steps": 8,
|
100 |
"gamma": 0.99,
|
101 |
"gae_lambda": 0.9,
|
a2c-AntBulletEnv-v0/policy.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 56190
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:68e57ac9d209d3c3a88634c2f8aac64d0de2e29309c2f03d540bb2a852b73da1
|
3 |
size 56190
|
a2c-AntBulletEnv-v0/policy.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 56958
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3b56c54511169ea81467c1efd02f69fc747ec61220815b80f83358ff4430de9c
|
3 |
size 56958
|
config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f7aa3f31ea0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f7aa3f31f30>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f7aa3f31fc0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f7aa3f32050>", "_build": "<function ActorCriticPolicy._build at 0x7f7aa3f320e0>", "forward": "<function ActorCriticPolicy.forward at 0x7f7aa3f32170>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f7aa3f32200>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f7aa3f32290>", "_predict": "<function ActorCriticPolicy._predict at 0x7f7aa3f32320>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f7aa3f323b0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f7aa3f32440>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f7aa3f324d0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f7aa3f38cc0>"}, "verbose": 1, "policy_kwargs": {":type:": "<class 'dict'>", ":serialized:": "gAWVowAAAAAAAAB9lCiMDGxvZ19zdGRfaW5pdJRK/v///4wKb3J0aG9faW5pdJSJjA9vcHRpbWl6ZXJfY2xhc3OUjBN0b3JjaC5vcHRpbS5ybXNwcm9wlIwHUk1TcHJvcJSTlIwQb3B0aW1pemVyX2t3YXJnc5R9lCiMBWFscGhhlEc/764UeuFHrowDZXBzlEc+5Pi1iONo8YwMd2VpZ2h0X2RlY2F5lEsAdXUu", "log_std_init": -2, "ortho_init": false, "optimizer_class": "<class 'torch.optim.rmsprop.RMSprop'>", "optimizer_kwargs": {"alpha": 0.99, "eps": 1e-05, "weight_decay": 0}}, "observation_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVZwIAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLHIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWcAAAAAAAAAAAAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/lGgKSxyFlIwBQ5R0lFKUjARoaWdolGgSKJZwAAAAAAAAAAAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH+UaApLHIWUaBV0lFKUjA1ib3VuZGVkX2JlbG93lGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLHIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaCFLHIWUaBV0lFKUjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "_shape": [28], "low": "[-inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf\n -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf]", "high": "[inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf\n inf inf inf inf inf inf inf inf inf inf]", "bounded_below": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "bounded_above": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "_np_random": null}, "action_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVnwEAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLCIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWIAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAv5RoCksIhZSMAUOUdJRSlIwEaGlnaJRoEiiWIAAAAAAAAAAAAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAP5RoCksIhZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolggAAAAAAAAAAQEBAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLCIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYIAAAAAAAAAAEBAQEBAQEBlGghSwiFlGgVdJRSlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "_shape": [8], "low": "[-1. -1. -1. -1. -1. -1. -1. -1.]", "high": "[1. 1. 1. 1. 1. 1. 1. 1.]", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_np_random": null}, "n_envs": 4, "num_timesteps": 10000000, "_total_timesteps": 10000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1676814694069137676, "learning_rate": 0.00096, "tensorboard_log": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS9vcHQvY29uZGEvbGliL3B5dGhvbjMuMTAvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuCQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvb3B0L2NvbmRhL2xpYi9weXRob24zLjEwL3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz9PdRBNVR1phZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAMoQnr8ohoA/kYx+va5Mfj8ApxHAXs+Jv1tRVb9t914++SC9Pk7DpD/bSIG/3JoPv86Gj784qvQ75ENxP4LNbzxAzz6/0MBOPypNxD3uoB6+MnefP3S3jrtvb/i+Cmq6vEiScj84e9y/1Du9Pv6m7b+ycg4+TegVP/Nczj6MDiG9KxSEPgRwPD/E5+8+MoZ4vk8olz8lN9S73FqPP3BmOD+SWOG+aA6iPmf66r5UMDy+Yh+GvuIApb60suk+L/z/PGxHZj99uw2/t8mTPsy2QT5IknI/zJ4UP9Q7vT7R4Qk/1TNGPjhq4r5pilA/P8B2P2Ygwz9IBRc/tB8DPxEyxj31KZc/dr7wu3LMXD+ZP1S/lBj3Pd/1IT+TzBC/bf/NvRinJT/coJK+KLHnPuLRnL4hzDK/Kz5Avx5y4r64JtC9AxaHv8yeFD/UO70+0eEJP2CGLb0j/Ly+ZC1OP6qVKr/5ENW+xUmfPc35dz3vUsq+YzSXP3ZEnbtPbBc/vaDhvNNPar+ieAo8Wj2kvUc4lTwavok/O92xPITv7D73Kww930UivqBJrDsMhh8+basOvUiScj/MnhQ/1Du9PtHhCT+UjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYEAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwSFlIwBQ5R0lFKULg=="}, "_last_original_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAAAAAD1eSG3AACAPwAAAAAAAAAAAAAAAAAAAAAAAACARfOtPQAAAACQCAHAAAAAAMYfCb4AAAAADtvmPwAAAAAARrK9AAAAAK+q9T8AAAAABQrlPQAAAADfnuq/AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAvsmstgAAgD8AAAAAAAAAAAAAAAAAAAAAAAAAgG6NkT0AAAAAhxXfvwAAAAC3vxi9AAAAACDy2j8AAAAAWmQQvgAAAAA2WPw/AAAAANlf/rwAAAAAfAvrvwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAEhKsrYAAIA/AAAAAAAAAAAAAAAAAAAAAAAAAIAjRoo7AAAAAOCW8r8AAAAA054TvQAAAAC07/Q/AAAAADT3mL0AAAAANwr5PwAAAADpfdo9AAAAAIw/3b8AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAADqPGa2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACAnFGGPQAAAAAS9OG/AAAAAFp7Br4AAAAAkwz/PwAAAAA9Z9a6AAAAAMAO2T8AAAAAZLUzvQAAAAA+4e6/AAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_episode_num": 0, "use_sde": true, "sde_sample_freq": -1, "_current_progress_remaining": 0.0, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVQwwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQIqeq2c8TzyMAWyUTegDjAF0lEdAx7eIyhSLqHV9lChoBkdAi8G0jLSuyWgHTegDaAhHQMe39eCTUy51fZQoaAZHQIfE3pdKNAFoB03oA2gIR0DHuHORPoFFdX2UKGgGR0CNqmUwBYFJaAdN6ANoCEdAx7nMPXCj13V9lChoBkdAiydS08eS0WgHTegDaAhHQMe55KJuVHF1fZQoaAZHQIKT3okiUxFoB03oA2gIR0DHulI2VE/jdX2UKGgGR0COcXhx5s0paAdN6ANoCEdAx7rO/20zCXV9lChoBkdAgR5iRnvlVGgHTegDaAhHQMe8RiNsFdN1fZQoaAZHQIMnn3i704BoB03oA2gIR0DHvF00aZQYdX2UKGgGR0CMwGySFGoaaAdN6ANoCEdAx7zabc45tHV9lChoBkdAjZ6+t8uzyGgHTegDaAhHQMe9V17IDHR1fZQoaAZHQId8xFmWdEtoB03oA2gIR0DHvq4FotcwdX2UKGgGR0CDiFk3CKrJaAdN6ANoCEdAx77GuMdcS3V9lChoBkdAjABBSDRMOGgHTegDaAhHQMe/MuSwGGF1fZQoaAZHwEWKGt6ol2NoB0uEaAhHQMe/gfLcKw91fZQoaAZHQIvR7544ZMtoB03oA2gIR0DHv6+oHcDbdX2UKGgGR0CDE+zFdcB2aAdN6ANoCEdAx8EEaWom5XV9lChoBkdAjeerThHby2gHTegDaAhHQMfBHMyBTXJ1fZQoaAZHQIg3DG96C19oB03oA2gIR0DHwdk2cawVdX2UKGgGR0CHzcKO1fE5aAdN6ANoCEdAx8IHnscABHV9lChoBkdAg5wljd56dGgHTegDaAhHQMfDYsuOCGx1fZQoaAZHQIeKIQSSNfhoB03oA2gIR0DHw3rnkkrxdX2UKGgGR0CCvkH/LkjpaAdN6ANoCEdAx8Q4sySFG3V9lChoBkdAgjkPRiPQwGgHTegDaAhHQMfEZzQNTcZ1fZQoaAZHQIW3emvW6LBoB03oA2gIR0DHxcAHTqjadX2UKGgGR0CJf5glWwNcaAdN6ANoCEdAx8XYBRyfc3V9lChoBkdAiCWsi0OVgWgHTegDaAhHQMfGk4BFNL11fZQoaAZHQIDtUw+MZP5oB03oA2gIR0DHxsL9l2/0dX2UKGgGR0CIpLpFCswMaAdN6ANoCEdAx8gYHPeHi3V9lChoBkdAhzlGxdIGyGgHTegDaAhHQMfIMCgK4QV1fZQoaAZHQIqUaad+XqtoB03oA2gIR0DHyOuLcbiqdX2UKGgGR0CFSMsNDtw8aAdN6ANoCEdAx8kZsUqQR3V9lChoBkdAgdKt9x6v7mgHTegDaAhHQMfKcUYj0MB1fZQoaAZHQIbnx8+iaiNoB03oA2gIR0DHyokIAwPAdX2UKGgGR0CF05QaaTfSaAdN6ANoCEdAx8tFAckt3HV9lChoBkdAht3RcE/0NGgHTegDaAhHQMfLc91dPcl1fZQoaAZHQIh0dDD0lJJoB03oA2gIR0DHzNELUkOadX2UKGgGR0CC6zBomG/OaAdN6ANoCEdAx8zpqPfbbnV9lChoBkdAeXB6lLvkR2gHTegDaAhHQMfNovNNahZ1fZQoaAZHQIZ/5Y1YQrdoB03oA2gIR0DHzdJFCswMdX2UKGgGR0CFxFIWgvlEaAdN6ANoCEdAx89HzCk43nV9lChoBkdAgpBJMHryD2gHTegDaAhHQMfPa/N7jT91fZQoaAZHQIiy9xXGOuJoB03oA2gIR0DH0C17rs0IdX2UKGgGR0CGqSSBbwBpaAdN6ANoCEdAx9BcAp8WsXV9lChoBkdAhuLvvrnkk2gHTegDaAhHQMfRuI7FKkF1fZQoaAZHQIPqnDej2zxoB03oA2gIR0DH0dEmtyPudX2UKGgGR0CG7D779AHFaAdN6ANoCEdAx9KQzYVZcXV9lChoBkdAhjRemNzbOGgHTegDaAhHQMfSv7ah6B11fZQoaAZHQIgZ27YkE9toB03oA2gIR0DH1BdJtix3dX2UKGgGR0CE8DguRLbpaAdN6ANoCEdAx9Qv1K5CnnV9lChoBkdAg6CpvYODrmgHTegDaAhHQMfU7YL1EmZ1fZQoaAZHQIdeVhVlwtJoB03oA2gIR0DH1RzSVnmJdX2UKGgGR0CH79pt78ekaAdN6ANoCEdAx9Z3SG8Em3V9lChoBkdAhCMLp7kXDWgHTegDaAhHQMfWkRB/qgR1fZQoaAZHQIQ2gk1Mue1oB03oA2gIR0DH101ELH+7dX2UKGgGR0CHXhPM0P6LaAdN6ANoCEdAx9d75TqB3HV9lChoBkdAhveln7Hhj2gHTegDaAhHQMfY2C704BF1fZQoaAZHQIRVsSuhbnpoB03oA2gIR0DH2PCw2VFAdX2UKGgGR0CGBfmQr+YMaAdN6ANoCEdAx9msNdZ7onV9lChoBkdAiYkLFOwgT2gHTegDaAhHQMfZ2yGSIP91fZQoaAZHQIv4ZU3n6mBoB03oA2gIR0DH2zODBdledX2UKGgGR0CJaAmuTzNEaAdN6ANoCEdAx9tLfHggo3V9lChoBkdAfIZjiXIEKWgHTegDaAhHQMfcC1sUIs11fZQoaAZHQIh1mhh6SkloB03oA2gIR0DH3Do8hcJMdX2UKGgGR0CIeJH9WIXTaAdN6ANoCEdAx92YLYPGyXV9lChoBkdAcwHt2LYPG2gHTegDaAhHQMfdsIuf29N1fZQoaAZHQIr0J4fOlftoB03oA2gIR0DH3m2E25xzdX2UKGgGR0B5ZPxb0OEvaAdN6ANoCEdAx96cNp/PPnV9lChoBkdAg2XiBwuM/GgHTegDaAhHQMff84t6HCZ1fZQoaAZHQIHisLORkmRoB03oA2gIR0DH4Aq4axX5dX2UKGgGR0CJprBl+VkdaAdN6ANoCEdAx+DXUpd8iXV9lChoBkdAhlN1zIV/MGgHTegDaAhHQMfhA+YtxuN1fZQoaAZHQIZp5sKsuFpoB03oA2gIR0DH4n1R1oxpdX2UKGgGR0B4pGVHFxXGaAdN6ANoCEdAx+KV0I1LrXV9lChoBkdAiMp77TDwY2gHTegDaAhHQMfjUwuuiex1fZQoaAZHQHqqAHeJpFloB03oA2gIR0DH44GvOhTPdX2UKGgGR0CBk7bjcVQAaAdN6ANoCEdAx+TeTA31jHV9lChoBkdAh+Ba8xsVL2gHTegDaAhHQMfk9rsKLKp1fZQoaAZHQIYMuC/XXiBoB03oA2gIR0DH5bUS5AhTdX2UKGgGR0CC6TgDRtxdaAdN6ANoCEdAx+XkrZrYXnV9lChoBkdAiDWet0V8C2gHTegDaAhHQMfnQZG8VYZ1fZQoaAZHQHazAkTpPh1oB03oA2gIR0DH51o9xIatdX2UKGgGR0CFYeZsKsuGaAdN6ANoCEdAx+gYJIDoyXV9lChoBkdAhd784YJmd2gHTegDaAhHQMfoRyV4X411fZQoaAZHQIaAc6NlyzZoB03oA2gIR0DH6aLoW56MdX2UKGgGR0CFVdx8UmD2aAdN6ANoCEdAx+m7RYRuj3V9lChoBkdAe8j5prULD2gHTegDaAhHQMfqfYnndO91fZQoaAZHQIYBZ8F6iTNoB03oA2gIR0DH6qvzWf9QdX2UKGgGR0CFWT1nuiN9aAdN6ANoCEdAx+wKtvn8sXV9lChoBkdAhZwfZdv862gHTegDaAhHQMfsI2i+L3t1fZQoaAZHQIw/PwVj7Q9oB03oA2gIR0DH7OL1dxACdX2UKGgGR0CFNW1tO2y+aAdN6ANoCEdAx+0R4FiazHV9lChoBkdAhzl76pHZsmgHTegDaAhHQMfucUvoNd91fZQoaAZHQIkwEsUZeiVoB03oA2gIR0DH7ooyEcsEdX2UKGgGR0CDd1DE3sHCaAdN6ANoCEdAx+9Kv3ai9XV9lChoBkdAh4EZle4TbmgHTegDaAhHQMfvefoRqXZ1fZQoaAZHQIVN+10DEFZoB03oA2gIR0DH8Nk12q1gdX2UKGgGR0CFhBzNliBoaAdN6ANoCEdAx/Dxq1PWQXV9lChoBkdAe+EoP07KaGgHTegDaAhHQMfxr9DIBBB1fZQoaAZHQIFInXNC7btoB03oA2gIR0DH8d/6O5rhdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 312500, "n_steps": 8, "gamma": 0.99, "gae_lambda": 0.9, "ent_coef": 0.0, "vf_coef": 0.4, "max_grad_norm": 0.5, "normalize_advantage": false, "system_info": {"OS": "Linux-5.4.0-137-generic-x86_64-with-glibc2.27 # 154-Ubuntu SMP Thu Jan 5 17:03:22 UTC 2023", "Python": "3.10.8", "Stable-Baselines3": "1.7.0", "PyTorch": "1.13.1", "GPU Enabled": "True", "Numpy": "1.22.3", "Gym": "0.21.0"}}
|
|
|
1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f9efa3f1ea0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f9efa3f1f30>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f9efa3f1fc0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f9efa3f2050>", "_build": "<function ActorCriticPolicy._build at 0x7f9efa3f20e0>", "forward": "<function ActorCriticPolicy.forward at 0x7f9efa3f2170>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f9efa3f2200>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f9efa3f2290>", "_predict": "<function ActorCriticPolicy._predict at 0x7f9efa3f2320>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f9efa3f23b0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f9efa3f2440>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f9efa3f24d0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f9efa3f8700>"}, "verbose": 1, "policy_kwargs": {":type:": "<class 'dict'>", ":serialized:": "gAWVowAAAAAAAAB9lCiMDGxvZ19zdGRfaW5pdJRK/v///4wKb3J0aG9faW5pdJSJjA9vcHRpbWl6ZXJfY2xhc3OUjBN0b3JjaC5vcHRpbS5ybXNwcm9wlIwHUk1TcHJvcJSTlIwQb3B0aW1pemVyX2t3YXJnc5R9lCiMBWFscGhhlEc/764UeuFHrowDZXBzlEc+5Pi1iONo8YwMd2VpZ2h0X2RlY2F5lEsAdXUu", "log_std_init": -2, "ortho_init": false, "optimizer_class": "<class 'torch.optim.rmsprop.RMSprop'>", "optimizer_kwargs": {"alpha": 0.99, "eps": 1e-05, "weight_decay": 0}}, "observation_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVZwIAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLHIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWcAAAAAAAAAAAAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/AACA/wAAgP8AAID/lGgKSxyFlIwBQ5R0lFKUjARoaWdolGgSKJZwAAAAAAAAAAAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH8AAIB/AACAfwAAgH+UaApLHIWUaBV0lFKUjA1ib3VuZGVkX2JlbG93lGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLHIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYcAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUaCFLHIWUaBV0lFKUjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "_shape": [28], "low": "[-inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf\n -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf -inf]", "high": "[inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf inf\n inf inf inf inf inf inf inf inf inf inf]", "bounded_below": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "bounded_above": "[False False False False False False False False False False False False\n False False False False False False False False False False False False\n False False False False]", "_np_random": null}, "action_space": {":type:": "<class 'gym.spaces.box.Box'>", ":serialized:": "gAWVnwEAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLCIWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWIAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAv5RoCksIhZSMAUOUdJRSlIwEaGlnaJRoEiiWIAAAAAAAAAAAAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAP5RoCksIhZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolggAAAAAAAAAAQEBAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLCIWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYIAAAAAAAAAAEBAQEBAQEBlGghSwiFlGgVdJRSlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "_shape": [8], "low": "[-1. -1. -1. -1. -1. -1. -1. -1.]", "high": "[1. 1. 1. 1. 1. 1. 1. 1.]", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_np_random": null}, "n_envs": 4, "num_timesteps": 2000000, "_total_timesteps": 2000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1676881276006055953, "learning_rate": 0.00096, "tensorboard_log": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS9vcHQvY29uZGEvbGliL3B5dGhvbjMuMTAvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuCQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvb3B0L2NvbmRhL2xpYi9weXRob24zLjEwL3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz9PdRBNVR1phZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAADoLK0Bn8gFAVV5AwMobp77OPhHAVT8MQFRfHEDT9ou/PkSRwNrTbT/cY8a+kbLqPp4a/T/Ziu2+Q6QHwJCoRz/xrRk/46w6vytRsz4bsIK/IC60Pt0tdsDkWARA/eurP6pEqr88xhY/ZdjmPoW9VD/LeDk/UW9pvyZ5yz5rAE8/ldoxv6mzRr9o+ok79hwpvz4+gj6uh7k+B/+8vjmgTMAi4xu/A+CmP81Cx70Kol4+odvzPiZ+7D+AF4E/s/KwvR28FL/PGrk9SWI4PzZ4TkCqRKq/1lTZv6jyDcA1B5q/O+LEP9HQib7RAgk/ZkkUP4kUCsAK7cs/HK4owIQDmb6yR4HAmTOLPl+AdT+rZ9M+yMOtv6FGVj187zk/gsLqvE1ECcBs9a2/vhoQwJS8gz+CWBS9cncCwJRSCT50nvO+BXNAPzzGFj9l2OY+NQeavyTChL/SYfi/JiELvzIzSr/n+RC/hN3uvsN4y7204O29/UYDP6a0br8oh8m++3p7v7ffxz4GrCS+2wMtPwgXZj4ToFs/8Tp8Pl5WFT/bLIq/FgRRP+uPtrv6ZTu+zT1GPgVzQD88xhY/ZdjmPjUHmr+UjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYEAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwSFlIwBQ5R0lFKULg=="}, "_last_original_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVNQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJbAAQAAAAAAAAAAAABMklW2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACAMKWyvQAAAAA5sOq/AAAAAHqagj0AAAAAxozvPwAAAAAJfgu+AAAAAATh4j8AAAAAm0YRPgAAAABDE/S/AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAX9zutgAAgD8AAAAAAAAAAAAAAAAAAAAAAAAAgBPUB7wAAAAAvxHrvwAAAAC+swC+AAAAACItAEAAAAAAjOZzvQAAAABtG+w/AAAAAFhUBD4AAAAA7+TkvwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAF3uNzYAAIA/AAAAAAAAAAAAAAAAAAAAAAAAAIAVhW49AAAAAH2q3r8AAAAAkpvuPQAAAAAsxfI/AAAAAF7bkjwAAAAAxID9PwAAAACHLBW9AAAAAFdV8r8AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAABQvsS2AACAPwAAAAAAAAAAAAAAAAAAAAAAAACA+C6kPQAAAAD0Wfy/AAAAAI2Sw7sAAAAAVTP2PwAAAABFVCK9AAAAANP68T8AAAAABFb9PQAAAACJE/e/AAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiSwRLHIaUjAFDlHSUUpQu"}, "_episode_num": 0, "use_sde": true, "sde_sample_freq": -1, "_current_progress_remaining": 0.0, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQIfbrqfOD8OMAWyUTegDjAF0lEdAoLlIDgZTAHV9lChoBkdAh98CYkVvdmgHTegDaAhHQKC6lUutfXx1fZQoaAZHQIsFS2hIvrZoB03oA2gIR0Cgu/GMOwxGdX2UKGgGR0CL0qVC5VfeaAdN6ANoCEdAoLwmZiNKiHV9lChoBkdAkvVLmZE2HmgHTegDaAhHQKDCKr6LwWp1fZQoaAZHQJOCgWP91lpoB03oA2gIR0Cgw3TrVvuPdX2UKGgGR0CNwraZhKDkaAdN6ANoCEdAoMTNkauOj3V9lChoBkdAlXBeEh7mdWgHTegDaAhHQKDFAjL0SRN1fZQoaAZHQJCLS6I3zc1oB03oA2gIR0CgyxmwJPZadX2UKGgGR0CKUy+mFajfaAdN6ANoCEdAoMxoaNuLrHV9lChoBkdAiAfRtxdY4mgHTegDaAhHQKDNvvIfbK11fZQoaAZHQIiqNb7j1f5oB03oA2gIR0CgzfM9KVY7dX2UKGgGR0CKMltzjm0WaAdN6ANoCEdAoNQdzbN8mnV9lChoBkdAiGUXE61b7mgHTegDaAhHQKDVdZowmE51fZQoaAZHQIR99FnZkCpoB03oA2gIR0Cg1tYXfqHHdX2UKGgGR0CPpIdNFjNIaAdN6ANoCEdAoNcLN6gM+nV9lChoBkdAh6l3OGCZnmgHTegDaAhHQKDdNiqhlDp1fZQoaAZHQIbMLCcf/3poB03oA2gIR0Cg3oTollbvdX2UKGgGR0CKityGzru6aAdN6ANoCEdAoN/eR7qptXV9lChoBkdAiuQXHaN+9mgHTegDaAhHQKDgE3uuzQh1fZQoaAZHQIoG4RTS9dxoB03oA2gIR0Cg5i4DTz/ZdX2UKGgGR0CLX2u/UONHaAdN6ANoCEdAoOd5/G2kSHV9lChoBkdAhxRiCz1K5GgHTegDaAhHQKDo0hQm/nJ1fZQoaAZHQIrVI9V3ljpoB03oA2gIR0Cg6QcZLqUvdX2UKGgGR0CKq9Ys/Y8MaAdN6ANoCEdAoO8cDB/I83V9lChoBkdAjZzgUUO/cmgHTegDaAhHQKDwZLeQ+2V1fZQoaAZHQJJ4/a24NI9oB03oA2gIR0Cg8b1GTcIrdX2UKGgGR0CM/+mtyPuHaAdN6ANoCEdAoPHxkI5YHXV9lChoBkdAjXyYMnZ00WgHTegDaAhHQKD4CpiqhlF1fZQoaAZHQIkhS9kBjnVoB03oA2gIR0Cg+VOoo/iYdX2UKGgGR0CKGxg4OtnxaAdN6ANoCEdAoPqvrY5DJHV9lChoBkdAiRPGsmv4d2gHTegDaAhHQKD65MV1wHZ1fZQoaAZHQIk2DKvFFUhoB03oA2gIR0ChAP1SwW30dX2UKGgGR0CQqFIMSbpeaAdN6ANoCEdAoQJHpdKNAHV9lChoBkdAh40Zk9U0emgHTegDaAhHQKEDnnFHavl1fZQoaAZHQI+v26TW5H5oB03oA2gIR0ChA9Mmv4dqdX2UKGgGR0CF12tmthd/aAdN6ANoCEdAoQnp15jYqXV9lChoBkdAkOmCzcAR02gHTegDaAhHQKELM/BWPtF1fZQoaAZHQIkNRiy6cy5oB03oA2gIR0ChDI3EyckMdX2UKGgGR0CQBtk1Mue0aAdN6ANoCEdAoQzCBRQ793V9lChoBkdAhkxcVYZEUmgHTegDaAhHQKES3kuHvc91fZQoaAZHQIYodqUNayNoB03oA2gIR0ChFCaX8fmtdX2UKGgGR0CTj4HJcPe6aAdN6ANoCEdAoRV9r2xptnV9lChoBkdAjixbLt/nXGgHTegDaAhHQKEVsSs8xKx1fZQoaAZHQIw617rs0HhoB03oA2gIR0ChG8eBH09RdX2UKGgGR0CLuXLPldTpaAdN6ANoCEdAoR0UeMhounV9lChoBkdAhcYrHEMspWgHTegDaAhHQKEecFK02Lp1fZQoaAZHQIl21/OMVDdoB03oA2gIR0ChHqTvJA+qdX2UKGgGR0CTGTsRxtHhaAdN6ANoCEdAoSS3BguyvHV9lChoBkdAinqWrfcesGgHTegDaAhHQKEmAq3mV7h1fZQoaAZHQJMFKUr08NhoB03oA2gIR0ChJ1l9jPOZdX2UKGgGR0CLY/a9sabXaAdN6ANoCEdAoSeOLpA2RHV9lChoBkdAk/ZG43FUAGgHTegDaAhHQKEtjqqwQlN1fZQoaAZHQJKYPEuQIUtoB03oA2gIR0ChLtpZOi35dX2UKGgGR0CQW7qYJE6UaAdN6ANoCEdAoTAxDNQj2XV9lChoBkdAk33pqREF4mgHTegDaAhHQKEwZLrX18N1fZQoaAZHQJHhiU7jkuJoB03oA2gIR0ChNmdNN8E3dX2UKGgGR0CRVX1L8JlbaAdN6ANoCEdAoTev6TGHYnV9lChoBkdAklCS+UQkHGgHTegDaAhHQKE5BrAP/aR1fZQoaAZHQJHnxhvze41oB03oA2gIR0ChOTtQKrq/dX2UKGgGR0CThMMbWEsbaAdN6ANoCEdAoT9Fa8pTdnV9lChoBkdAkP6HW8RL9WgHTegDaAhHQKFAj8rqdH51fZQoaAZHQI6Adg4OtnxoB03oA2gIR0ChQeSH2ys0dX2UKGgGR0CVMUDEFW4maAdN6ANoCEdAoUIY7xNIsnV9lChoBkdAkzjmn0kGA2gHTegDaAhHQKFIFqrzXjF1fZQoaAZHQJTvoRChN/RoB03oA2gIR0ChSV0jLSuydX2UKGgGR0CPWpyWAwwkaAdN6ANoCEdAoUq1qL0jDHV9lChoBkdAlLlv/NqxkmgHTegDaAhHQKFK6eQMhHN1fZQoaAZHQJR1CGpMpPRoB03oA2gIR0ChUPdznzQNdX2UKGgGR0CR8LvCdjG2aAdN6ANoCEdAoVJAgxJumHV9lChoBkdAlHzGJzkp7WgHTegDaAhHQKFTlszEaVF1fZQoaAZHQJQu73ztkWhoB03oA2gIR0ChU8sefZmJdX2UKGgGR0CTc3phF3INaAdN6ANoCEdAoVneMbWEsnV9lChoBkdAlYFFHFxXGWgHTegDaAhHQKFbKCz1K5F1fZQoaAZHQJEcmy2QXANoB03oA2gIR0ChXH9weeWfdX2UKGgGR0CMs9M5fdAPaAdN6ANoCEdAoVyzyDqW1XV9lChoBkdAkcuoHPeHi2gHTegDaAhHQKFivCvX9R91fZQoaAZHQInX+uRs/INoB03oA2gIR0ChZASaEzwddX2UKGgGR0CUHVc/+sHTaAdN6ANoCEdAoWVcMiKR+3V9lChoBkdAkNCZMQEpzGgHTegDaAhHQKFlkHKOktV1fZQoaAZHQI8zDjYI0IloB03oA2gIR0Cha5hhYvFndX2UKGgGR0CTI5i704BFaAdN6ANoCEdAoWzehXbM5nV9lChoBkdAkke7R8c+7mgHTegDaAhHQKFuMmsNlRR1fZQoaAZHQJSvLqoqCpZoB03oA2gIR0ChbmY0VJtjdX2UKGgGR0CWEztj0+TvaAdN6ANoCEdAoXRj6ciGFnV9lChoBkdAlHwgzP8htGgHTegDaAhHQKF1p1loUSJ1fZQoaAZHQJODk8nuy/toB03oA2gIR0ChdvuxrzoVdX2UKGgGR0CUL6LRa5f/aAdN6ANoCEdAoXcv3Dej23V9lChoBkdAit3q6nR9gGgHTegDaAhHQKF9Pe7+T/11fZQoaAZHQJSQTKB/ZuhoB03oA2gIR0ChfoSOaOPvdX2UKGgGR0CR4gU+LWI5aAdN6ANoCEdAoX/Zjvuw5nV9lChoBkdAktoDF+/gzmgHTegDaAhHQKGADORkmQd1fZQoaAZHQJbUj9ZRsM1oB03oA2gIR0Chhg0oa1kUdX2UKGgGR0CWbi44p+c6aAdN6ANoCEdAoYdT3PAwf3V9lChoBkdAjL4/IsAeaWgHTegDaAhHQKGIqKjzqbB1fZQoaAZHQJKcaxGDtgNoB03oA2gIR0ChiNyB06o3dX2UKGgGR0CWHQmbb1yvaAdN6ANoCEdAoY7ULv1DjXV9lChoBkdAk8OYK+i8F2gHTegDaAhHQKGQF0se4kN1fZQoaAZHQJWHovIwM6RoB03oA2gIR0ChkWkzfrKOdX2UKGgGR0CSPfom5UcXaAdN6ANoCEdAoZGcewLVnXVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 62500, "n_steps": 8, "gamma": 0.99, "gae_lambda": 0.9, "ent_coef": 0.0, "vf_coef": 0.4, "max_grad_norm": 0.5, "normalize_advantage": false, "system_info": {"OS": "Linux-5.4.0-137-generic-x86_64-with-glibc2.27 # 154-Ubuntu SMP Thu Jan 5 17:03:22 UTC 2023", "Python": "3.10.8", "Stable-Baselines3": "1.7.0", "PyTorch": "1.13.1", "GPU Enabled": "True", "Numpy": "1.22.3", "Gym": "0.21.0"}}
|
replay.mp4
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66b4b378f2ce782886ad26574de152bc0f5a95496bf928a3d3e43986aae28219
|
3 |
+
size 1089677
|
results.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"mean_reward":
|
|
|
1 |
+
{"mean_reward": 859.1235057426417, "std_reward": 326.9194550238592, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-02-20T08:59:10.861816"}
|