enaitzb commited on
Commit
d2a5a71
1 Parent(s): 59b62ac

Initial commit

Browse files
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: stable-baselines3
3
+ tags:
4
+ - PandaPush-v3
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - stable-baselines3
8
+ model-index:
9
+ - name: TQC
10
+ results:
11
+ - task:
12
+ type: reinforcement-learning
13
+ name: reinforcement-learning
14
+ dataset:
15
+ name: PandaPush-v3
16
+ type: PandaPush-v3
17
+ metrics:
18
+ - type: mean_reward
19
+ value: -45.00 +/- 15.00
20
+ name: mean_reward
21
+ verified: false
22
+ ---
23
+
24
+ # **TQC** Agent playing **PandaPush-v3**
25
+ This is a trained model of a **TQC** agent playing **PandaPush-v3**
26
+ using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
27
+
28
+ ## Usage (with Stable-baselines3)
29
+ TODO: Add your code
30
+
31
+
32
+ ```python
33
+ from stable_baselines3 import ...
34
+ from huggingface_sb3 import load_from_hub
35
+
36
+ ...
37
+ ```
TQC-PandaPush-v3.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:641241c9d629ff249437275da22b65c7d94014c5b959e1655c2b727e99290013
3
+ size 24237541
TQC-PandaPush-v3/_stable_baselines3_version ADDED
@@ -0,0 +1 @@
 
 
1
+ 2.2.1
TQC-PandaPush-v3/actor.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df6d9762cccb9023d471fdf1885adc94c8ff5193a94c37d6d2b76855015ab2b6
3
+ size 4337724
TQC-PandaPush-v3/critic.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac003d64bee25b05f24534e66386daf821ca16ec0036ce6911913ced2205bf02
3
+ size 8852358
TQC-PandaPush-v3/data ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "policy_class": {
3
+ ":type:": "<class 'abc.ABCMeta'>",
4
+ ":serialized:": "gAWVMQAAAAAAAACMGHNiM19jb250cmliLnRxYy5wb2xpY2llc5SMEE11bHRpSW5wdXRQb2xpY3mUk5Qu",
5
+ "__module__": "sb3_contrib.tqc.policies",
6
+ "__doc__": "\n Policy class (with both actor and critic) for TQC.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param use_expln: Use ``expln()`` function instead of ``exp()`` when using gSDE to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param clip_mean: Clip the mean output when using gSDE to avoid numerical instability.\n :param features_extractor_class: Features extractor to use.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_quantiles: Number of quantiles for the critic.\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ",
7
+ "__init__": "<function MultiInputPolicy.__init__ at 0x7f675132b5b0>",
8
+ "__abstractmethods__": "frozenset()",
9
+ "_abc_impl": "<_abc._abc_data object at 0x7f6751332f40>"
10
+ },
11
+ "verbose": 1,
12
+ "policy_kwargs": {
13
+ "net_arch": [
14
+ 512,
15
+ 512,
16
+ 512
17
+ ],
18
+ "n_critics": 2,
19
+ "use_sde": false
20
+ },
21
+ "num_timesteps": 125008,
22
+ "_total_timesteps": 125000,
23
+ "_num_timesteps_at_start": 0,
24
+ "seed": null,
25
+ "action_noise": null,
26
+ "start_time": 1704880180866482285,
27
+ "learning_rate": 0.001,
28
+ "tensorboard_log": "./TQC/logs",
29
+ "_last_obs": null,
30
+ "_last_episode_starts": {
31
+ ":type:": "<class 'numpy.ndarray'>",
32
+ ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAEBAQEBAQEBAQEBAQEBAQGUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="
33
+ },
34
+ "_last_original_obs": {
35
+ ":type:": "<class 'collections.OrderedDict'>",
36
+ ":serialized:": "gAWV+wYAAAAAAACMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojA1hY2hpZXZlZF9nb2FslIwSbnVtcHkuY29yZS5udW1lcmljlIwLX2Zyb21idWZmZXKUk5QolsAAAAAAAAAAJwnuPeZ2kzsswaM8LyAmPtipZT2kSqM8WSZNvkfYjz6AYaM8JtSCvZiSZr12waM8as2bPEUy4T0K16M8p7+ZPIkBSr7HwaM8SmJ1vKV89b3iv6M85q1rO+0waD2DSa48Aol1vjK+iz4hwqM8h+0KvoWjc70RwqM8sg5VPg4V6LxxwaM8u2tnPXiwaz0K16M8c8dAPp7Tkb0MwqM8LNAevkHdxr3IwaM8C/pWvuvevb5zncK+nvSvPeltSLzp5KI8lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwOGlIwBQ5R0lFKUjAxkZXNpcmVkX2dvYWyUaAcolsAAAAAAAAAArbLnPWwZ4b0K16M8rmWjvYzMED4K16M8GOzPvb7U9T0K16M8oI8HPoAhoj0K16M8dHEjPTkx2r0K16M8WUXGPewnrr0K16M8JDeMPXAdhD0K16M8YMeyvaTPmb0K16M80jr4vXn+uz0K16M8P3iAvaqMlz0K16M8LgejPfF0VL0K16M8oqIxvd/0h7wK16M858IQPgju/70K16M8jeEDvtH3+DwK16M8F+rOPRNME74K16M8je4OPujL0b0K16M8lGgOSxBLA4aUaBJ0lFKUjAtvYnNlcnZhdGlvbpRoByiWgAQAAAAAAAARtek9+4gsPU1pnT029Ru+GE7BvlaFur8nCe495naTOyzBozzZasUuFQqzt7qOMrcTt5s3MvzhMSuqmzenKrS0tE9zOscTibEb/Q89/atdPcQykDvxvOo+osyKP8E9Wb0vICY+2KllPaRKozxdj4229Sc6u20goz+U4xk/cO7iPhU8Ijv3Pjq+JrelO6nHqUBKaR++qWSdPmwtZjvwh/0+Ajkiv4hSgrxZJk2+R9iPPoBhozxIeRW7o26lt2PFIL+rPBg8etqavcMIAzvS4qY9AIZpvRBihcBLvvm9XF8XviEQrTy2xw4/6BtyP72ul78m1IK9mJJmvXbBozyks1SxOOJyt21ts7eUM1M3pi5HsWhAUzdRPxc0BgElOtFFEbHqch09GWwarEMjSj4AAAAAAAAAgAAAAABqzZs8RTLhPQrXozwAAAAAAAAAgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACeJye7IuGwvRk7Dj7YmYe+LXdDv3102r6nv5k8iQFKvsfBozwXl2w6/kJwukcPuz6bMQE+Zti0vgxEcze1bBs+aW6NvSZQtj6PVAe9wDn3vepPBT596Zu+TaIJv5Ogjb9KYnW8pXz1veK/ozyPSYk4ADgbN44XULcqtKm3xgV7tuJTxrVW45o4NO25MiSD+7haQQQ9MwzPPbBmbTzOr6K+ckBavSOJFD/mrWs77TBoPYNJrjxaO/o6UemRve1t+j6QyTm/2luBvq7VfjtTUJW9mIYhvQmTn0GF1Ny8UkdnPpJfKj16o56+E5TfPlS4pb4CiXW+Mr6LPiHCozylilO1342TtpHoCMA1UwG/8Q+nPhucRziJu8O6xGoIOxqjxsBccj6+Gge0vXNYFj2rkL8+y6/dPtT8nD6H7Qq+haNzvRHCozwAAAAA2w/JP1ERuL9qOAezhQsVs9TLjTMA4ug1JkjTtcoEvqggN/U9r6O0vfwEBj1IrNM8mA3fPhzjnr6yDlU+DhXovHHBozyFLUMz3IR8N5SyNMCFdEo380+DNt6BVDcfsE+5x7cdOkIFPDbqch09GWwarEMjSj4AAAAAAAAAgAAAAAC7a2c9eLBrPQrXozwAAAAAAAAAgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAADDP+k9kz89vZ0rGj2kwre+EBJDP1nhOb9zx0A+ntORvQzCozyoJVw0nIWxNAdgvL7kxI60rko/NGgivzQ6KQe3rl5ct1w4QzPvKTK90b11vmEzpju1ZD4+ZO8wP06p6rws0B6+Qd3GvcjBozzxfmU39P6gNnrPAUDzPEG2S3cDthDpF61pIqyvu08IMOTeczhHFlS+LUsZvzAuDz64cXu/wiJLv75+Pj8L+la+6969vnOdwr5YfMc/K3zIv5RxXj1inqi+DYdpOuLBizs5IS88cBpJvnqKa714K/89qiwjPdVprzvuGwk+LsXkPSJaOr2e9K896W1IvOnkojxzAbe6rgB5u+e3Y79fL9c9sDwUv+3oHLw4Eb6+1gLMviggVcGUaA5LEEsShpRoEnSUUpR1Lg==",
37
+ "achieved_goal": "[[ 0.11622839 0.00450026 0.01998957]\n [ 0.16223215 0.05607018 0.01993305]\n [-0.2003416 0.28094694 0.01994395]\n [-0.0638812 -0.05629215 0.01998971]\n [ 0.01901885 0.10995916 0.02 ]\n [ 0.01876814 -0.19727148 0.01998986]\n [-0.01497705 -0.11986665 0.01998896]\n [ 0.00359618 0.05668728 0.02127529]\n [-0.23978046 0.27293545 0.01999003]\n [-0.13567172 -0.05948212 0.01999 ]\n [ 0.20806387 -0.02833035 0.0199897 ]\n [ 0.05649922 0.05754134 0.02 ]\n [ 0.18826084 -0.07120441 0.01998999]\n [-0.15509099 -0.0971017 0.01998986]\n [-0.20993821 -0.37084135 -0.3801075 ]\n [ 0.08591579 -0.01223324 0.01988454]]",
38
+ "desired_goal": "[[ 0.11313377 -0.10991177 0.02 ]\n [-0.07978378 0.14140528 0.02 ]\n [-0.10152453 0.12003468 0.02 ]\n [ 0.13238382 0.07916546 0.02 ]\n [ 0.03990312 -0.1065392 0.02 ]\n [ 0.09681196 -0.08503708 0.02 ]\n [ 0.06846455 0.06450927 0.02 ]\n [-0.08729434 -0.07510307 0.02 ]\n [-0.12120594 0.09179396 0.02 ]\n [-0.06272935 0.07399876 0.02 ]\n [ 0.07960354 -0.05186934 0.02 ]\n [-0.04336799 -0.01659626 0.02 ]\n [ 0.1413685 -0.12496573 0.02 ]\n [-0.1287901 0.03039161 0.02 ]\n [ 0.10103243 -0.14384489 0.02 ]\n [ 0.13958187 -0.1024397 0.02 ]]",
39
+ "observation": "[[ 1.14114888e-01 4.21228223e-02 7.68610016e-02 -1.52302593e-01\n -3.77548933e-01 -1.45719409e+00 1.16228394e-01 4.50025778e-03\n 1.99895725e-02 8.97750277e-11 -2.13431576e-05 -1.06428579e-05\n 1.85626905e-05 6.57703314e-09 1.85566805e-05 -3.35586464e-07\n 9.28159105e-04 -3.98947320e-09]\n [ 3.51534896e-02 5.41190989e-02 4.40058298e-03 4.58472759e-01\n 1.08436990e+00 -5.30374087e-02 1.62232146e-01 5.60701787e-02\n 1.99330524e-02 -4.21881714e-06 -2.84051639e-03 1.27442706e+00\n 6.01128817e-01 4.43225384e-01 2.47550500e-03 -1.81880817e-01\n 5.05723339e-03 5.30562258e+00]\n [-1.55675083e-01 3.07408601e-01 3.51222884e-03 4.95177746e-01\n -6.33682370e-01 -1.59084946e-02 -2.00341597e-01 2.80946940e-01\n 1.99439526e-02 -2.28078850e-03 -1.97210520e-05 -6.28011882e-01\n 9.29180812e-03 -7.56120235e-02 1.99942361e-03 8.14873129e-02\n -5.70125580e-02 -4.16822052e+00]\n [-1.21944986e-01 -1.47824705e-01 2.11258549e-02 5.57734847e-01\n 9.45738316e-01 -1.18502009e+00 -6.38812035e-02 -5.62921464e-02\n 1.99897103e-02 -3.09521742e-09 -1.44769947e-05 -2.13894182e-05\n 1.25885890e-05 -2.89848279e-09 1.25915758e-05 1.40860053e-07\n 6.29440299e-04 -2.11399631e-09]\n [ 3.84396687e-02 -2.19447225e-12 1.97400138e-01 0.00000000e+00\n -0.00000000e+00 0.00000000e+00 1.90188475e-02 1.09959163e-01\n 1.99999996e-02 0.00000000e+00 -0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00 0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00]\n [-2.55057914e-03 -8.63669068e-02 1.38897315e-01 -2.64845610e-01\n -7.63537228e-01 -4.26669985e-01 1.87681448e-02 -1.97271481e-01\n 1.99898612e-02 9.02519969e-04 -9.16525605e-04 3.65350932e-01\n 1.26165792e-01 -3.53213489e-01 1.44997721e-05 1.51781872e-01\n -6.90582469e-02 3.56080234e-01]\n [-3.30396257e-02 -1.20715618e-01 1.30187660e-01 -3.04515749e-01\n -5.37632763e-01 -1.10646284e+00 -1.49770472e-02 -1.19866647e-01\n 1.99889578e-02 6.54637042e-05 9.25175846e-06 -1.24032504e-05\n -2.02302654e-05 -3.74052752e-06 -1.47765627e-06 7.38563685e-05\n 2.16447020e-08 -1.19930395e-04]\n [ 3.22888866e-02 1.01097487e-01 1.44898146e-02 -3.17747533e-01\n -5.32841161e-02 5.80217540e-01 3.59618058e-03 5.66872843e-02\n 2.12752875e-02 1.90911745e-03 -7.12457970e-02 4.89119917e-01\n -7.25731850e-01 -2.52653897e-01 3.88846872e-03 -7.29071125e-02\n -3.94349992e-02 1.99467945e+01]\n [-2.69568060e-02 2.25858003e-01 4.15950492e-02 -3.09840977e-01\n 4.36676592e-01 -3.23671937e-01 -2.39780456e-01 2.72935450e-01\n 1.99900288e-02 -7.88053796e-07 -4.39745736e-06 -2.13919473e+00\n -5.05175889e-01 3.26293498e-01 4.75906818e-05 -1.49332092e-03\n 2.08155904e-03 -6.20740986e+00]\n [-1.85983121e-01 -8.79041702e-02 3.67054455e-02 3.74150604e-01\n 4.32981819e-01 3.06616426e-01 -1.35671720e-01 -5.94821163e-02\n 1.99899990e-02 0.00000000e+00 1.57079637e+00 -1.43802845e+00\n -3.14834452e-08 -3.47022429e-08 6.60290027e-08 1.73511216e-06\n -1.57417230e-06 -2.10963144e-14]\n [ 1.19734049e-01 -8.82028267e-02 3.27195972e-02 2.58389860e-02\n 4.35650587e-01 -3.10326457e-01 2.08063871e-01 -2.83303522e-02\n 1.99897010e-02 4.54433753e-08 1.50513042e-05 -2.82339954e+00\n 1.20672676e-05 3.91341155e-06 1.26664218e-05 -1.98066686e-04\n 6.01645967e-04 2.80172435e-06]\n [ 3.84396687e-02 -2.19447225e-12 1.97400138e-01 0.00000000e+00\n -0.00000000e+00 0.00000000e+00 5.64992242e-02 5.75413406e-02\n 1.99999996e-02 0.00000000e+00 -0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00 0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00]\n [ 1.13891147e-01 -4.62032072e-02 3.76392491e-02 -3.58906865e-01\n 7.61994362e-01 -7.26094782e-01 1.88260838e-01 -7.12044090e-02\n 1.99899897e-02 2.05027959e-07 3.30660328e-07 -3.67920130e-01\n -2.65928179e-07 1.78154295e-07 3.56015562e-07 -8.05622585e-06\n -1.31350662e-05 4.54532341e-08]\n [-4.34970222e-02 -2.39981905e-01 5.07204281e-03 1.85931042e-01\n 6.91152811e-01 -2.86451839e-02 -1.55090988e-01 -9.71016958e-02\n 1.99898630e-02 1.36790195e-05 4.79805203e-06 2.02828836e+00\n -2.87947182e-06 -1.95899588e-06 -8.63510652e-12 -3.13110676e-10\n 4.95898156e-10 5.81432978e-05]\n [-2.07116231e-01 -5.98803341e-01 1.39824629e-01 -9.82203960e-01\n -7.93499112e-01 7.44121432e-01 -2.09938213e-01 -3.70841354e-01\n -3.80107492e-01 1.55848217e+00 -1.56628931e+00 5.43075353e-02\n -3.29333365e-01 8.90836876e-04 4.26505599e-03 1.06890732e-02\n -1.96389914e-01 -5.75051084e-02]\n [ 1.24594629e-01 3.98375168e-02 5.35319233e-03 1.33895606e-01\n 1.11704215e-01 -4.54961136e-02 8.59157890e-02 -1.22332359e-02\n 1.98845435e-02 -1.39622239e-03 -3.79947899e-03 -8.89524877e-01\n 1.05070822e-01 -5.79051018e-01 -9.57701821e-03 -3.71225119e-01\n -3.98459136e-01 -1.33203506e+01]]"
40
+ },
41
+ "_episode_num": 4894,
42
+ "use_sde": false,
43
+ "sde_sample_freq": -1,
44
+ "_current_progress_remaining": -6.4000000000064e-05,
45
+ "_stats_window_size": 100,
46
+ "ep_info_buffer": {
47
+ ":type:": "<class 'collections.deque'>",
48
+ ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHwCYAAAAAAACMAWyUSwyMAXSUR0C8fKkFGG21dX2UKGgGR8AoAAAAAAAAaAdLDWgIR0C8e0v3evZAdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fV2WUr08dX2UKGgGR8A2AAAAAAAAaAdLF2gIR0C8ffCwSrYHdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8e+Wxlg+hdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fdVZLZi/dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fSQntv4udX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8fZZtSAH3dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fpIpDu0DdX2UKGgGR8AuAAAAAAAAaAdLEGgIR0C8fZxhhH9WdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8f4tcbBGhdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8f0eAuqWDdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8f3icTakAdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8gHU1ZTybdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8gaklu3tsdX2UKGgGR8BDgAAAAAAAaAdLKGgIR0C8gUHeFcptdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8gzKcEvCedX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8gt6UJOWTdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8g2qvicXndX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8hU/UvwmWdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8honqmj0udX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8hmg2ETQFdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8iApA+pwTdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8iHDPOY6XdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8iQkfYBeYdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8iGVYZEUkdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8ics9B8hLdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8iWLPhQ3xdX2UKGgGR8A6AAAAAAAAaAdLG2gIR0C8iLYw22ofdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8iq655JK8dX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8ikNuUD+zdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8i6YYFaB7dX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8irKZlWfcdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8jREvGp++dX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8i9rEYO2BdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8jc4MKCxvdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8jG+evpyIdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8jszJMg2ZdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8jdhU3n6mdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8jlVpKzzFdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8jra/M4cWdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8kQ9zjm0WdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8kMsLjPv8dX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8kW0I9kjHdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8kYFhXr+pdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8kG+0G/vfdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8kjlNL128dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8kvDr/sE8dX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8lUJ/G2kSdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8lqAiiZfEdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8la5BPbfxdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8lrUWykbhdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8lb9ELH+7dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8l7F/MGHIdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8l22P91lodX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8lnaF7D2rdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8mCOocaOxdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8l9+loDgZdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8ljuIl+mWdX2UKGgGR8AuAAAAAAAAaAdLEGgIR0C8ls4aDPGAdX2UKGgGR8AwAAAAAAAAaAdLEWgIR0C8mZXI6r/9dX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8mMxhH9WIdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8mfCJwbVCdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8mXAIIF/ydX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8mgFHe7+UdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8m47GWD6FdX2UKGgGR8AsAAAAAAAAaAdLD2gIR0C8mg+HJtBOdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8mxm51/2CdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8m0IZ/CqIdX2UKGgGR8AqAAAAAAAAaAdLDmgIR0C8nlvcN6PbdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8ndyhi9ZidX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8nx1nmJWOdX2UKGgGR8A6AAAAAAAAaAdLG2gIR0C8nm10tAcDdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8n0ijDbaidX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8np8kUsWgdX2UKGgGR8AuAAAAAAAAaAdLEGgIR0C8n/khV2iddX2UKGgGR8A2AAAAAAAAaAdLF2gIR0C8oRKGHpKSdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8oKkIsyzpdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8oh/bsWwedX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8pQFeBxxUdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8pS1K9PDYdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8pzVpsXSCdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8pkHb212JdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8prTVhCtzdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8qFN3jdYXdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8qz1EiMYNdX2UKGgGR8AzAAAAAAAAaAdLFGgIR0C8qrSl3yI6dX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8qknGS6lMdX2UKGgGR8A5AAAAAAAAaAdLGmgIR0C8quJ/LDAKdX2UKGgGR8AsAAAAAAAAaAdLD2gIR0C8q8TKPn0TdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8q3vJiiItdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8rvp4bCJodX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8rmySJTESdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8rVcp1A7gdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8r5ntKIzndX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8sPLpNbkfdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8sCLDqGDddX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8sZFbu+h5dX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8sh/3nIQwdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8sS1M/QjVdWUu"
49
+ },
50
+ "ep_success_buffer": {
51
+ ":type:": "<class 'collections.deque'>",
52
+ ":serialized:": "gAWVhgAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKIiIiYiIiYmIiYiIiIiIiIiIiYiIiYiIiImJiIiIiIiIiIiIiIiIiIiIiImIiIiIiYiIiYmJiYmIiIiIiIiIiImIiIiIiIiJiIiIiIiIiIiIiYiIiIiIiIiIiIiJiImIiYiIiIllLg=="
53
+ },
54
+ "_n_updates": 249824,
55
+ "observation_space": {
56
+ ":type:": "<class 'gymnasium.spaces.dict.Dict'>",
57
+ ":serialized:": "gAWVKAQAAAAAAACMFWd5bW5hc2l1bS5zcGFjZXMuZGljdJSMBERpY3SUk5QpgZR9lCiMBnNwYWNlc5SMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojA1hY2hpZXZlZF9nb2FslIwUZ3ltbmFzaXVtLnNwYWNlcy5ib3iUjANCb3iUk5QpgZR9lCiMBWR0eXBllIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYowNYm91bmRlZF9iZWxvd5SMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYDAAAAAAAAAAEBAZRoE4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksDhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoHCiWAwAAAAAAAAABAQGUaCBLA4WUaCR0lFKUjAZfc2hhcGWUSwOFlIwDbG93lGgcKJYMAAAAAAAAAAAAIMEAACDBAAAgwZRoFksDhZRoJHSUUpSMBGhpZ2iUaBwolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgWSwOFlGgkdJRSlIwIbG93X3JlcHKUjAUtMTAuMJSMCWhpZ2hfcmVwcpSMBDEwLjCUjApfbnBfcmFuZG9tlE51YowMZGVzaXJlZF9nb2FslGgNKYGUfZQoaBBoFmgZaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgnaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgsSwOFlGguaBwolgwAAAAAAAAAAAAgwQAAIMEAACDBlGgWSwOFlGgkdJRSlGgzaBwolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgWSwOFlGgkdJRSlGg4jAUtMTAuMJRoOowEMTAuMJRoPE51YowLb2JzZXJ2YXRpb26UaA0pgZR9lChoEGgWaBloHCiWEgAAAAAAAAABAQEBAQEBAQEBAQEBAQEBAQGUaCBLEoWUaCR0lFKUaCdoHCiWEgAAAAAAAAABAQEBAQEBAQEBAQEBAQEBAQGUaCBLEoWUaCR0lFKUaCxLEoWUaC5oHCiWSAAAAAAAAAAAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMGUaBZLEoWUaCR0lFKUaDNoHCiWSAAAAAAAAAAAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEGUaBZLEoWUaCR0lFKUaDiMBS0xMC4wlGg6jAQxMC4wlGg8TnVidWgsTmgQTmg8TnViLg==",
58
+ "spaces": "OrderedDict([('achieved_goal', Box(-10.0, 10.0, (3,), float32)), ('desired_goal', Box(-10.0, 10.0, (3,), float32)), ('observation', Box(-10.0, 10.0, (18,), float32))])",
59
+ "_shape": null,
60
+ "dtype": null,
61
+ "_np_random": null
62
+ },
63
+ "action_space": {
64
+ ":type:": "<class 'gymnasium.spaces.box.Box'>",
65
+ ":serialized:": "gAWVRgIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWAwAAAAAAAAABAQGUaAiMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLA4WUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBEolgMAAAAAAAAAAQEBlGgVSwOFlGgZdJRSlIwGX3NoYXBllEsDhZSMA2xvd5RoESiWDAAAAAAAAAAAAIC/AACAvwAAgL+UaAtLA4WUaBl0lFKUjARoaWdolGgRKJYMAAAAAAAAAAAAgD8AAIA/AACAP5RoC0sDhZRoGXSUUpSMCGxvd19yZXBylIwELTEuMJSMCWhpZ2hfcmVwcpSMAzEuMJSMCl9ucF9yYW5kb22UjBRudW1weS5yYW5kb20uX3BpY2tsZZSMEF9fZ2VuZXJhdG9yX2N0b3KUk5SMBVBDRzY0lIWUUpR9lCiMDWJpdF9nZW5lcmF0b3KUjAVQQ0c2NJSMBXN0YXRllH2UKGg7ihD2oi47LX+jhyYPVDe4PGw9jANpbmOUihEhcLX2BvXSlZ6u2JdzbZmLAHWMCmhhc191aW50MzKUSwCMCHVpbnRlZ2VylEsAdWJ1Yi4=",
66
+ "dtype": "float32",
67
+ "bounded_below": "[ True True True]",
68
+ "bounded_above": "[ True True True]",
69
+ "_shape": [
70
+ 3
71
+ ],
72
+ "low": "[-1. -1. -1.]",
73
+ "high": "[1. 1. 1.]",
74
+ "low_repr": "-1.0",
75
+ "high_repr": "1.0",
76
+ "_np_random": "Generator(PCG64)"
77
+ },
78
+ "n_envs": 1,
79
+ "buffer_size": 1000000,
80
+ "batch_size": 2048,
81
+ "learning_starts": 100,
82
+ "tau": 0.05,
83
+ "gamma": 0.95,
84
+ "gradient_steps": -1,
85
+ "optimize_memory_usage": false,
86
+ "replay_buffer_class": {
87
+ ":type:": "<class 'abc.ABCMeta'>",
88
+ ":serialized:": "gAWVPwAAAAAAAACMJ3N0YWJsZV9iYXNlbGluZXMzLmhlci5oZXJfcmVwbGF5X2J1ZmZlcpSMD0hlclJlcGxheUJ1ZmZlcpSTlC4=",
89
+ "__module__": "stable_baselines3.her.her_replay_buffer",
90
+ "__annotations__": "{'env': typing.Optional[stable_baselines3.common.vec_env.base_vec_env.VecEnv]}",
91
+ "__doc__": "\n Hindsight Experience Replay (HER) buffer.\n Paper: https://arxiv.org/abs/1707.01495\n\n Replay buffer for sampling HER (Hindsight Experience Replay) transitions.\n\n .. note::\n\n Compared to other implementations, the ``future`` goal sampling strategy is inclusive:\n the current transition can be used when re-sampling.\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param env: The training environment\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n Disabled for now (see https://github.com/DLR-RM/stable-baselines3/pull/243#discussion_r531535702)\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n :param n_sampled_goal: Number of virtual transitions to create per real transition,\n by sampling new goals.\n :param goal_selection_strategy: Strategy for sampling goals for replay.\n One of ['episode', 'final', 'future']\n :param copy_info_dict: Whether to copy the info dictionary and pass it to\n ``compute_reward()`` method.\n Please note that the copy may cause a slowdown.\n False by default.\n ",
92
+ "__init__": "<function HerReplayBuffer.__init__ at 0x7f67512c31c0>",
93
+ "__getstate__": "<function HerReplayBuffer.__getstate__ at 0x7f67512c3250>",
94
+ "__setstate__": "<function HerReplayBuffer.__setstate__ at 0x7f67512c32e0>",
95
+ "set_env": "<function HerReplayBuffer.set_env at 0x7f67512c3370>",
96
+ "add": "<function HerReplayBuffer.add at 0x7f67512c3400>",
97
+ "_compute_episode_length": "<function HerReplayBuffer._compute_episode_length at 0x7f67512c3490>",
98
+ "sample": "<function HerReplayBuffer.sample at 0x7f67512c3520>",
99
+ "_get_real_samples": "<function HerReplayBuffer._get_real_samples at 0x7f67512c35b0>",
100
+ "_get_virtual_samples": "<function HerReplayBuffer._get_virtual_samples at 0x7f67512c3640>",
101
+ "_sample_goals": "<function HerReplayBuffer._sample_goals at 0x7f67512c36d0>",
102
+ "truncate_last_trajectory": "<function HerReplayBuffer.truncate_last_trajectory at 0x7f67512c3760>",
103
+ "__abstractmethods__": "frozenset()",
104
+ "_abc_impl": "<_abc._abc_data object at 0x7f67512d2400>"
105
+ },
106
+ "replay_buffer_kwargs": {
107
+ "goal_selection_strategy": "future",
108
+ "n_sampled_goal": 4
109
+ },
110
+ "train_freq": {
111
+ ":type:": "<class 'stable_baselines3.common.type_aliases.TrainFreq'>",
112
+ ":serialized:": "gAWVYQAAAAAAAACMJXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi50eXBlX2FsaWFzZXOUjAlUcmFpbkZyZXGUk5RLAWgAjBJUcmFpbkZyZXF1ZW5jeVVuaXSUk5SMBHN0ZXCUhZRSlIaUgZQu"
113
+ },
114
+ "use_sde_at_warmup": false,
115
+ "target_entropy": -3.0,
116
+ "ent_coef": "auto",
117
+ "target_update_interval": 1,
118
+ "top_quantiles_to_drop_per_net": 2,
119
+ "lr_schedule": {
120
+ ":type:": "<class 'function'>",
121
+ ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuDQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz9QYk3S8an8hZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"
122
+ },
123
+ "batch_norm_stats": [],
124
+ "batch_norm_stats_target": []
125
+ }
TQC-PandaPush-v3/ent_coef_optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c82aaee374f961336bce557a9a9e8ac323c2d98baadfb2cca1dfb1784c7364e
3
+ size 1940
TQC-PandaPush-v3/policy.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ead807d5c3805cd1e7235c450141ab3aa1db2c7827f4d3aacb1a2a694ddf6686
3
+ size 11019096
TQC-PandaPush-v3/pytorch_variables.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d74631f18384de55ed4cd8f7f21a5a74ad3a7a7e661130c6d51ec4935d11637
3
+ size 1180
TQC-PandaPush-v3/system_info.txt ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - OS: Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023
2
+ - Python: 3.10.12
3
+ - Stable-Baselines3: 2.2.1
4
+ - PyTorch: 2.1.0+cu121
5
+ - GPU Enabled: False
6
+ - Numpy: 1.23.5
7
+ - Cloudpickle: 2.2.1
8
+ - Gymnasium: 0.29.1
9
+ - OpenAI Gym: 0.25.2
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVMQAAAAAAAACMGHNiM19jb250cmliLnRxYy5wb2xpY2llc5SMEE11bHRpSW5wdXRQb2xpY3mUk5Qu", "__module__": "sb3_contrib.tqc.policies", "__doc__": "\n Policy class (with both actor and critic) for TQC.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param use_expln: Use ``expln()`` function instead of ``exp()`` when using gSDE to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param clip_mean: Clip the mean output when using gSDE to avoid numerical instability.\n :param features_extractor_class: Features extractor to use.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_quantiles: Number of quantiles for the critic.\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ", "__init__": "<function MultiInputPolicy.__init__ at 0x7f675132b5b0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f6751332f40>"}, "verbose": 1, "policy_kwargs": {"net_arch": [512, 512, 512], "n_critics": 2, "use_sde": false}, "num_timesteps": 125008, "_total_timesteps": 125000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1704880180866482285, "learning_rate": 0.001, "tensorboard_log": "./TQC/logs", "_last_obs": null, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAEBAQEBAQEBAQEBAQEBAQGUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": {":type:": "<class 'collections.OrderedDict'>", ":serialized:": "gAWV+wYAAAAAAACMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojA1hY2hpZXZlZF9nb2FslIwSbnVtcHkuY29yZS5udW1lcmljlIwLX2Zyb21idWZmZXKUk5QolsAAAAAAAAAAJwnuPeZ2kzsswaM8LyAmPtipZT2kSqM8WSZNvkfYjz6AYaM8JtSCvZiSZr12waM8as2bPEUy4T0K16M8p7+ZPIkBSr7HwaM8SmJ1vKV89b3iv6M85q1rO+0waD2DSa48Aol1vjK+iz4hwqM8h+0KvoWjc70RwqM8sg5VPg4V6LxxwaM8u2tnPXiwaz0K16M8c8dAPp7Tkb0MwqM8LNAevkHdxr3IwaM8C/pWvuvevb5zncK+nvSvPeltSLzp5KI8lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwOGlIwBQ5R0lFKUjAxkZXNpcmVkX2dvYWyUaAcolsAAAAAAAAAArbLnPWwZ4b0K16M8rmWjvYzMED4K16M8GOzPvb7U9T0K16M8oI8HPoAhoj0K16M8dHEjPTkx2r0K16M8WUXGPewnrr0K16M8JDeMPXAdhD0K16M8YMeyvaTPmb0K16M80jr4vXn+uz0K16M8P3iAvaqMlz0K16M8LgejPfF0VL0K16M8oqIxvd/0h7wK16M858IQPgju/70K16M8jeEDvtH3+DwK16M8F+rOPRNME74K16M8je4OPujL0b0K16M8lGgOSxBLA4aUaBJ0lFKUjAtvYnNlcnZhdGlvbpRoByiWgAQAAAAAAAARtek9+4gsPU1pnT029Ru+GE7BvlaFur8nCe495naTOyzBozzZasUuFQqzt7qOMrcTt5s3MvzhMSuqmzenKrS0tE9zOscTibEb/Q89/atdPcQykDvxvOo+osyKP8E9Wb0vICY+2KllPaRKozxdj4229Sc6u20goz+U4xk/cO7iPhU8Ijv3Pjq+JrelO6nHqUBKaR++qWSdPmwtZjvwh/0+Ajkiv4hSgrxZJk2+R9iPPoBhozxIeRW7o26lt2PFIL+rPBg8etqavcMIAzvS4qY9AIZpvRBihcBLvvm9XF8XviEQrTy2xw4/6BtyP72ul78m1IK9mJJmvXbBozyks1SxOOJyt21ts7eUM1M3pi5HsWhAUzdRPxc0BgElOtFFEbHqch09GWwarEMjSj4AAAAAAAAAgAAAAABqzZs8RTLhPQrXozwAAAAAAAAAgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACeJye7IuGwvRk7Dj7YmYe+LXdDv3102r6nv5k8iQFKvsfBozwXl2w6/kJwukcPuz6bMQE+Zti0vgxEcze1bBs+aW6NvSZQtj6PVAe9wDn3vepPBT596Zu+TaIJv5Ogjb9KYnW8pXz1veK/ozyPSYk4ADgbN44XULcqtKm3xgV7tuJTxrVW45o4NO25MiSD+7haQQQ9MwzPPbBmbTzOr6K+ckBavSOJFD/mrWs77TBoPYNJrjxaO/o6UemRve1t+j6QyTm/2luBvq7VfjtTUJW9mIYhvQmTn0GF1Ny8UkdnPpJfKj16o56+E5TfPlS4pb4CiXW+Mr6LPiHCozylilO1342TtpHoCMA1UwG/8Q+nPhucRziJu8O6xGoIOxqjxsBccj6+Gge0vXNYFj2rkL8+y6/dPtT8nD6H7Qq+haNzvRHCozwAAAAA2w/JP1ERuL9qOAezhQsVs9TLjTMA4ug1JkjTtcoEvqggN/U9r6O0vfwEBj1IrNM8mA3fPhzjnr6yDlU+DhXovHHBozyFLUMz3IR8N5SyNMCFdEo380+DNt6BVDcfsE+5x7cdOkIFPDbqch09GWwarEMjSj4AAAAAAAAAgAAAAAC7a2c9eLBrPQrXozwAAAAAAAAAgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAADDP+k9kz89vZ0rGj2kwre+EBJDP1nhOb9zx0A+ntORvQzCozyoJVw0nIWxNAdgvL7kxI60rko/NGgivzQ6KQe3rl5ct1w4QzPvKTK90b11vmEzpju1ZD4+ZO8wP06p6rws0B6+Qd3GvcjBozzxfmU39P6gNnrPAUDzPEG2S3cDthDpF61pIqyvu08IMOTeczhHFlS+LUsZvzAuDz64cXu/wiJLv75+Pj8L+la+6969vnOdwr5YfMc/K3zIv5RxXj1inqi+DYdpOuLBizs5IS88cBpJvnqKa714K/89qiwjPdVprzvuGwk+LsXkPSJaOr2e9K896W1IvOnkojxzAbe6rgB5u+e3Y79fL9c9sDwUv+3oHLw4Eb6+1gLMviggVcGUaA5LEEsShpRoEnSUUpR1Lg==", "achieved_goal": "[[ 0.11622839 0.00450026 0.01998957]\n [ 0.16223215 0.05607018 0.01993305]\n [-0.2003416 0.28094694 0.01994395]\n [-0.0638812 -0.05629215 0.01998971]\n [ 0.01901885 0.10995916 0.02 ]\n [ 0.01876814 -0.19727148 0.01998986]\n [-0.01497705 -0.11986665 0.01998896]\n [ 0.00359618 0.05668728 0.02127529]\n [-0.23978046 0.27293545 0.01999003]\n [-0.13567172 -0.05948212 0.01999 ]\n [ 0.20806387 -0.02833035 0.0199897 ]\n [ 0.05649922 0.05754134 0.02 ]\n [ 0.18826084 -0.07120441 0.01998999]\n [-0.15509099 -0.0971017 0.01998986]\n [-0.20993821 -0.37084135 -0.3801075 ]\n [ 0.08591579 -0.01223324 0.01988454]]", "desired_goal": "[[ 0.11313377 -0.10991177 0.02 ]\n [-0.07978378 0.14140528 0.02 ]\n [-0.10152453 0.12003468 0.02 ]\n [ 0.13238382 0.07916546 0.02 ]\n [ 0.03990312 -0.1065392 0.02 ]\n [ 0.09681196 -0.08503708 0.02 ]\n [ 0.06846455 0.06450927 0.02 ]\n [-0.08729434 -0.07510307 0.02 ]\n [-0.12120594 0.09179396 0.02 ]\n [-0.06272935 0.07399876 0.02 ]\n [ 0.07960354 -0.05186934 0.02 ]\n [-0.04336799 -0.01659626 0.02 ]\n [ 0.1413685 -0.12496573 0.02 ]\n [-0.1287901 0.03039161 0.02 ]\n [ 0.10103243 -0.14384489 0.02 ]\n [ 0.13958187 -0.1024397 0.02 ]]", "observation": "[[ 1.14114888e-01 4.21228223e-02 7.68610016e-02 -1.52302593e-01\n -3.77548933e-01 -1.45719409e+00 1.16228394e-01 4.50025778e-03\n 1.99895725e-02 8.97750277e-11 -2.13431576e-05 -1.06428579e-05\n 1.85626905e-05 6.57703314e-09 1.85566805e-05 -3.35586464e-07\n 9.28159105e-04 -3.98947320e-09]\n [ 3.51534896e-02 5.41190989e-02 4.40058298e-03 4.58472759e-01\n 1.08436990e+00 -5.30374087e-02 1.62232146e-01 5.60701787e-02\n 1.99330524e-02 -4.21881714e-06 -2.84051639e-03 1.27442706e+00\n 6.01128817e-01 4.43225384e-01 2.47550500e-03 -1.81880817e-01\n 5.05723339e-03 5.30562258e+00]\n [-1.55675083e-01 3.07408601e-01 3.51222884e-03 4.95177746e-01\n -6.33682370e-01 -1.59084946e-02 -2.00341597e-01 2.80946940e-01\n 1.99439526e-02 -2.28078850e-03 -1.97210520e-05 -6.28011882e-01\n 9.29180812e-03 -7.56120235e-02 1.99942361e-03 8.14873129e-02\n -5.70125580e-02 -4.16822052e+00]\n [-1.21944986e-01 -1.47824705e-01 2.11258549e-02 5.57734847e-01\n 9.45738316e-01 -1.18502009e+00 -6.38812035e-02 -5.62921464e-02\n 1.99897103e-02 -3.09521742e-09 -1.44769947e-05 -2.13894182e-05\n 1.25885890e-05 -2.89848279e-09 1.25915758e-05 1.40860053e-07\n 6.29440299e-04 -2.11399631e-09]\n [ 3.84396687e-02 -2.19447225e-12 1.97400138e-01 0.00000000e+00\n -0.00000000e+00 0.00000000e+00 1.90188475e-02 1.09959163e-01\n 1.99999996e-02 0.00000000e+00 -0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00 0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00]\n [-2.55057914e-03 -8.63669068e-02 1.38897315e-01 -2.64845610e-01\n -7.63537228e-01 -4.26669985e-01 1.87681448e-02 -1.97271481e-01\n 1.99898612e-02 9.02519969e-04 -9.16525605e-04 3.65350932e-01\n 1.26165792e-01 -3.53213489e-01 1.44997721e-05 1.51781872e-01\n -6.90582469e-02 3.56080234e-01]\n [-3.30396257e-02 -1.20715618e-01 1.30187660e-01 -3.04515749e-01\n -5.37632763e-01 -1.10646284e+00 -1.49770472e-02 -1.19866647e-01\n 1.99889578e-02 6.54637042e-05 9.25175846e-06 -1.24032504e-05\n -2.02302654e-05 -3.74052752e-06 -1.47765627e-06 7.38563685e-05\n 2.16447020e-08 -1.19930395e-04]\n [ 3.22888866e-02 1.01097487e-01 1.44898146e-02 -3.17747533e-01\n -5.32841161e-02 5.80217540e-01 3.59618058e-03 5.66872843e-02\n 2.12752875e-02 1.90911745e-03 -7.12457970e-02 4.89119917e-01\n -7.25731850e-01 -2.52653897e-01 3.88846872e-03 -7.29071125e-02\n -3.94349992e-02 1.99467945e+01]\n [-2.69568060e-02 2.25858003e-01 4.15950492e-02 -3.09840977e-01\n 4.36676592e-01 -3.23671937e-01 -2.39780456e-01 2.72935450e-01\n 1.99900288e-02 -7.88053796e-07 -4.39745736e-06 -2.13919473e+00\n -5.05175889e-01 3.26293498e-01 4.75906818e-05 -1.49332092e-03\n 2.08155904e-03 -6.20740986e+00]\n [-1.85983121e-01 -8.79041702e-02 3.67054455e-02 3.74150604e-01\n 4.32981819e-01 3.06616426e-01 -1.35671720e-01 -5.94821163e-02\n 1.99899990e-02 0.00000000e+00 1.57079637e+00 -1.43802845e+00\n -3.14834452e-08 -3.47022429e-08 6.60290027e-08 1.73511216e-06\n -1.57417230e-06 -2.10963144e-14]\n [ 1.19734049e-01 -8.82028267e-02 3.27195972e-02 2.58389860e-02\n 4.35650587e-01 -3.10326457e-01 2.08063871e-01 -2.83303522e-02\n 1.99897010e-02 4.54433753e-08 1.50513042e-05 -2.82339954e+00\n 1.20672676e-05 3.91341155e-06 1.26664218e-05 -1.98066686e-04\n 6.01645967e-04 2.80172435e-06]\n [ 3.84396687e-02 -2.19447225e-12 1.97400138e-01 0.00000000e+00\n -0.00000000e+00 0.00000000e+00 5.64992242e-02 5.75413406e-02\n 1.99999996e-02 0.00000000e+00 -0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00 0.00000000e+00 0.00000000e+00\n 0.00000000e+00 0.00000000e+00]\n [ 1.13891147e-01 -4.62032072e-02 3.76392491e-02 -3.58906865e-01\n 7.61994362e-01 -7.26094782e-01 1.88260838e-01 -7.12044090e-02\n 1.99899897e-02 2.05027959e-07 3.30660328e-07 -3.67920130e-01\n -2.65928179e-07 1.78154295e-07 3.56015562e-07 -8.05622585e-06\n -1.31350662e-05 4.54532341e-08]\n [-4.34970222e-02 -2.39981905e-01 5.07204281e-03 1.85931042e-01\n 6.91152811e-01 -2.86451839e-02 -1.55090988e-01 -9.71016958e-02\n 1.99898630e-02 1.36790195e-05 4.79805203e-06 2.02828836e+00\n -2.87947182e-06 -1.95899588e-06 -8.63510652e-12 -3.13110676e-10\n 4.95898156e-10 5.81432978e-05]\n [-2.07116231e-01 -5.98803341e-01 1.39824629e-01 -9.82203960e-01\n -7.93499112e-01 7.44121432e-01 -2.09938213e-01 -3.70841354e-01\n -3.80107492e-01 1.55848217e+00 -1.56628931e+00 5.43075353e-02\n -3.29333365e-01 8.90836876e-04 4.26505599e-03 1.06890732e-02\n -1.96389914e-01 -5.75051084e-02]\n [ 1.24594629e-01 3.98375168e-02 5.35319233e-03 1.33895606e-01\n 1.11704215e-01 -4.54961136e-02 8.59157890e-02 -1.22332359e-02\n 1.98845435e-02 -1.39622239e-03 -3.79947899e-03 -8.89524877e-01\n 1.05070822e-01 -5.79051018e-01 -9.57701821e-03 -3.71225119e-01\n -3.98459136e-01 -1.33203506e+01]]"}, "_episode_num": 4894, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -6.4000000000064e-05, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHwCYAAAAAAACMAWyUSwyMAXSUR0C8fKkFGG21dX2UKGgGR8AoAAAAAAAAaAdLDWgIR0C8e0v3evZAdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fV2WUr08dX2UKGgGR8A2AAAAAAAAaAdLF2gIR0C8ffCwSrYHdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8e+Wxlg+hdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fdVZLZi/dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fSQntv4udX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8fZZtSAH3dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8fpIpDu0DdX2UKGgGR8AuAAAAAAAAaAdLEGgIR0C8fZxhhH9WdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8f4tcbBGhdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8f0eAuqWDdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8f3icTakAdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8gHU1ZTybdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8gaklu3tsdX2UKGgGR8BDgAAAAAAAaAdLKGgIR0C8gUHeFcptdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8gzKcEvCedX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8gt6UJOWTdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8g2qvicXndX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8hU/UvwmWdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8honqmj0udX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8hmg2ETQFdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8iApA+pwTdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8iHDPOY6XdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8iQkfYBeYdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8iGVYZEUkdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8ics9B8hLdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8iWLPhQ3xdX2UKGgGR8A6AAAAAAAAaAdLG2gIR0C8iLYw22ofdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8iq655JK8dX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8ikNuUD+zdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8i6YYFaB7dX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8irKZlWfcdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8jREvGp++dX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8i9rEYO2BdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8jc4MKCxvdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8jG+evpyIdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8jszJMg2ZdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8jdhU3n6mdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8jlVpKzzFdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8jra/M4cWdX2UKGgGR8AQAAAAAAAAaAdLBWgIR0C8kQ9zjm0WdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8kMsLjPv8dX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8kW0I9kjHdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8kYFhXr+pdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8kG+0G/vfdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8kjlNL128dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8kvDr/sE8dX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8lUJ/G2kSdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8lqAiiZfEdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8la5BPbfxdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8lrUWykbhdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8lb9ELH+7dX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8l7F/MGHIdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8l22P91lodX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8lnaF7D2rdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8mCOocaOxdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8l9+loDgZdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8ljuIl+mWdX2UKGgGR8AuAAAAAAAAaAdLEGgIR0C8ls4aDPGAdX2UKGgGR8AwAAAAAAAAaAdLEWgIR0C8mZXI6r/9dX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8mMxhH9WIdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8mfCJwbVCdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8mXAIIF/ydX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8mgFHe7+UdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8m47GWD6FdX2UKGgGR8AsAAAAAAAAaAdLD2gIR0C8mg+HJtBOdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8mxm51/2CdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8m0IZ/CqIdX2UKGgGR8AqAAAAAAAAaAdLDmgIR0C8nlvcN6PbdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8ndyhi9ZidX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8nx1nmJWOdX2UKGgGR8A6AAAAAAAAaAdLG2gIR0C8nm10tAcDdX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8n0ijDbaidX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8np8kUsWgdX2UKGgGR8AuAAAAAAAAaAdLEGgIR0C8n/khV2iddX2UKGgGR8A2AAAAAAAAaAdLF2gIR0C8oRKGHpKSdX2UKGgGR8AUAAAAAAAAaAdLBmgIR0C8oKkIsyzpdX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8oh/bsWwedX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8pQFeBxxUdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8pS1K9PDYdX2UKGgGR8AiAAAAAAAAaAdLCmgIR0C8pzVpsXSCdX2UKGgGR8AkAAAAAAAAaAdLC2gIR0C8pkHb212JdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8prTVhCtzdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8qFN3jdYXdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8qz1EiMYNdX2UKGgGR8AzAAAAAAAAaAdLFGgIR0C8qrSl3yI6dX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8qknGS6lMdX2UKGgGR8A5AAAAAAAAaAdLGmgIR0C8quJ/LDAKdX2UKGgGR8AsAAAAAAAAaAdLD2gIR0C8q8TKPn0TdX2UKGgGR8AYAAAAAAAAaAdLB2gIR0C8q3vJiiItdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8rvp4bCJodX2UKGgGR8AcAAAAAAAAaAdLCGgIR0C8rmySJTESdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8rVcp1A7gdX2UKGgGR8AgAAAAAAAAaAdLCWgIR0C8r5ntKIzndX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8sPLpNbkfdX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8sCLDqGDddX2UKGgGR8AmAAAAAAAAaAdLDGgIR0C8sZFbu+h5dX2UKGgGRwAAAAAAAAAAaAdLAWgIR0C8sh/3nIQwdX2UKGgGR8BJAAAAAAAAaAdLMmgIR0C8sS1M/QjVdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVhgAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKIiIiYiIiYmIiYiIiIiIiIiIiYiIiYiIiImJiIiIiIiIiIiIiIiIiIiIiImIiIiIiYiIiYmJiYmIiIiIiIiIiImIiIiIiIiJiIiIiIiIiIiIiYiIiIiIiIiIiIiJiImIiYiIiIllLg=="}, "_n_updates": 249824, "observation_space": {":type:": "<class 'gymnasium.spaces.dict.Dict'>", ":serialized:": "gAWVKAQAAAAAAACMFWd5bW5hc2l1bS5zcGFjZXMuZGljdJSMBERpY3SUk5QpgZR9lCiMBnNwYWNlc5SMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojA1hY2hpZXZlZF9nb2FslIwUZ3ltbmFzaXVtLnNwYWNlcy5ib3iUjANCb3iUk5QpgZR9lCiMBWR0eXBllIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYowNYm91bmRlZF9iZWxvd5SMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYDAAAAAAAAAAEBAZRoE4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksDhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoHCiWAwAAAAAAAAABAQGUaCBLA4WUaCR0lFKUjAZfc2hhcGWUSwOFlIwDbG93lGgcKJYMAAAAAAAAAAAAIMEAACDBAAAgwZRoFksDhZRoJHSUUpSMBGhpZ2iUaBwolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgWSwOFlGgkdJRSlIwIbG93X3JlcHKUjAUtMTAuMJSMCWhpZ2hfcmVwcpSMBDEwLjCUjApfbnBfcmFuZG9tlE51YowMZGVzaXJlZF9nb2FslGgNKYGUfZQoaBBoFmgZaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgnaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgsSwOFlGguaBwolgwAAAAAAAAAAAAgwQAAIMEAACDBlGgWSwOFlGgkdJRSlGgzaBwolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgWSwOFlGgkdJRSlGg4jAUtMTAuMJRoOowEMTAuMJRoPE51YowLb2JzZXJ2YXRpb26UaA0pgZR9lChoEGgWaBloHCiWEgAAAAAAAAABAQEBAQEBAQEBAQEBAQEBAQGUaCBLEoWUaCR0lFKUaCdoHCiWEgAAAAAAAAABAQEBAQEBAQEBAQEBAQEBAQGUaCBLEoWUaCR0lFKUaCxLEoWUaC5oHCiWSAAAAAAAAAAAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMEAACDBAAAgwQAAIMGUaBZLEoWUaCR0lFKUaDNoHCiWSAAAAAAAAAAAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEEAACBBAAAgQQAAIEGUaBZLEoWUaCR0lFKUaDiMBS0xMC4wlGg6jAQxMC4wlGg8TnVidWgsTmgQTmg8TnViLg==", "spaces": "OrderedDict([('achieved_goal', Box(-10.0, 10.0, (3,), float32)), ('desired_goal', Box(-10.0, 10.0, (3,), float32)), ('observation', Box(-10.0, 10.0, (18,), float32))])", "_shape": null, "dtype": null, "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVRgIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWAwAAAAAAAAABAQGUaAiMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLA4WUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBEolgMAAAAAAAAAAQEBlGgVSwOFlGgZdJRSlIwGX3NoYXBllEsDhZSMA2xvd5RoESiWDAAAAAAAAAAAAIC/AACAvwAAgL+UaAtLA4WUaBl0lFKUjARoaWdolGgRKJYMAAAAAAAAAAAAgD8AAIA/AACAP5RoC0sDhZRoGXSUUpSMCGxvd19yZXBylIwELTEuMJSMCWhpZ2hfcmVwcpSMAzEuMJSMCl9ucF9yYW5kb22UjBRudW1weS5yYW5kb20uX3BpY2tsZZSMEF9fZ2VuZXJhdG9yX2N0b3KUk5SMBVBDRzY0lIWUUpR9lCiMDWJpdF9nZW5lcmF0b3KUjAVQQ0c2NJSMBXN0YXRllH2UKGg7ihD2oi47LX+jhyYPVDe4PGw9jANpbmOUihEhcLX2BvXSlZ6u2JdzbZmLAHWMCmhhc191aW50MzKUSwCMCHVpbnRlZ2VylEsAdWJ1Yi4=", "dtype": "float32", "bounded_below": "[ True True True]", "bounded_above": "[ True True True]", "_shape": [3], "low": "[-1. -1. -1.]", "high": "[1. 1. 1.]", "low_repr": "-1.0", "high_repr": "1.0", "_np_random": "Generator(PCG64)"}, "n_envs": 1, "buffer_size": 1000000, "batch_size": 2048, "learning_starts": 100, "tau": 0.05, "gamma": 0.95, "gradient_steps": -1, "optimize_memory_usage": false, "replay_buffer_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVPwAAAAAAAACMJ3N0YWJsZV9iYXNlbGluZXMzLmhlci5oZXJfcmVwbGF5X2J1ZmZlcpSMD0hlclJlcGxheUJ1ZmZlcpSTlC4=", "__module__": "stable_baselines3.her.her_replay_buffer", "__annotations__": "{'env': typing.Optional[stable_baselines3.common.vec_env.base_vec_env.VecEnv]}", "__doc__": "\n Hindsight Experience Replay (HER) buffer.\n Paper: https://arxiv.org/abs/1707.01495\n\n Replay buffer for sampling HER (Hindsight Experience Replay) transitions.\n\n .. note::\n\n Compared to other implementations, the ``future`` goal sampling strategy is inclusive:\n the current transition can be used when re-sampling.\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param env: The training environment\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n Disabled for now (see https://github.com/DLR-RM/stable-baselines3/pull/243#discussion_r531535702)\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n :param n_sampled_goal: Number of virtual transitions to create per real transition,\n by sampling new goals.\n :param goal_selection_strategy: Strategy for sampling goals for replay.\n One of ['episode', 'final', 'future']\n :param copy_info_dict: Whether to copy the info dictionary and pass it to\n ``compute_reward()`` method.\n Please note that the copy may cause a slowdown.\n False by default.\n ", "__init__": "<function HerReplayBuffer.__init__ at 0x7f67512c31c0>", "__getstate__": "<function HerReplayBuffer.__getstate__ at 0x7f67512c3250>", "__setstate__": "<function HerReplayBuffer.__setstate__ at 0x7f67512c32e0>", "set_env": "<function HerReplayBuffer.set_env at 0x7f67512c3370>", "add": "<function HerReplayBuffer.add at 0x7f67512c3400>", "_compute_episode_length": "<function HerReplayBuffer._compute_episode_length at 0x7f67512c3490>", "sample": "<function HerReplayBuffer.sample at 0x7f67512c3520>", "_get_real_samples": "<function HerReplayBuffer._get_real_samples at 0x7f67512c35b0>", "_get_virtual_samples": "<function HerReplayBuffer._get_virtual_samples at 0x7f67512c3640>", "_sample_goals": "<function HerReplayBuffer._sample_goals at 0x7f67512c36d0>", "truncate_last_trajectory": "<function HerReplayBuffer.truncate_last_trajectory at 0x7f67512c3760>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f67512d2400>"}, "replay_buffer_kwargs": {"goal_selection_strategy": "future", "n_sampled_goal": 4}, "train_freq": {":type:": "<class 'stable_baselines3.common.type_aliases.TrainFreq'>", ":serialized:": "gAWVYQAAAAAAAACMJXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi50eXBlX2FsaWFzZXOUjAlUcmFpbkZyZXGUk5RLAWgAjBJUcmFpbkZyZXF1ZW5jeVVuaXSUk5SMBHN0ZXCUhZRSlIaUgZQu"}, "use_sde_at_warmup": false, "target_entropy": -3.0, "ent_coef": "auto", "target_update_interval": 1, "top_quantiles_to_drop_per_net": 2, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuDQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz9QYk3S8an8hZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "batch_norm_stats": [], "batch_norm_stats_target": [], "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.2.1", "PyTorch": "2.1.0+cu121", "GPU Enabled": "False", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.29.1", "OpenAI Gym": "0.25.2"}}
replay.mp4 ADDED
Binary file (806 kB). View file
 
results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"mean_reward": -45.0, "std_reward": 15.0, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-01-10T14:54:41.604893"}