MadFritz commited on
Commit
bbd7bc2
·
1 Parent(s): dfec284

Upload td3 BipedalWalker-v3 trained agent

Browse files
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: stable-baselines3
3
+ tags:
4
+ - BipedalWalker-v3
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - stable-baselines3
8
+ model-index:
9
+ - name: td3
10
+ results:
11
+ - task:
12
+ type: reinforcement-learning
13
+ name: reinforcement-learning
14
+ dataset:
15
+ name: BipedalWalker-v3
16
+ type: BipedalWalker-v3
17
+ metrics:
18
+ - type: mean_reward
19
+ value: 306.79 +/- 1.20
20
+ name: mean_reward
21
+ verified: false
22
+ ---
23
+
24
+ # **td3** Agent playing **BipedalWalker-v3**
25
+ This is a trained model of a **td3** agent playing **BipedalWalker-v3**
26
+ using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
27
+
28
+ ## Usage (with Stable-baselines3)
29
+ TODO: Add your code
30
+
31
+
32
+ ```python
33
+ from stable_baselines3 import ...
34
+ from huggingface_sb3 import load_from_hub
35
+
36
+ ...
37
+ ```
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVMAAAAAAAAACMHnN0YWJsZV9iYXNlbGluZXMzLnRkMy5wb2xpY2llc5SMCVREM1BvbGljeZSTlC4=", "__module__": "stable_baselines3.td3.policies", "__annotations__": "{'actor': <class 'stable_baselines3.td3.policies.Actor'>, 'actor_target': <class 'stable_baselines3.td3.policies.Actor'>, 'critic': <class 'stable_baselines3.common.policies.ContinuousCritic'>, 'critic_target': <class 'stable_baselines3.common.policies.ContinuousCritic'>}", "__doc__": "\n Policy class (with both actor and critic) for TD3.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ", "__init__": "<function TD3Policy.__init__ at 0x0000020A8CE6D7E0>", "_build": "<function TD3Policy._build at 0x0000020A8CE6D870>", "_get_constructor_parameters": "<function TD3Policy._get_constructor_parameters at 0x0000020A8CE6D900>", "make_actor": "<function TD3Policy.make_actor at 0x0000020A8CE6D990>", "make_critic": "<function TD3Policy.make_critic at 0x0000020A8CE6DA20>", "forward": "<function TD3Policy.forward at 0x0000020A8CE6DAB0>", "_predict": "<function TD3Policy._predict at 0x0000020A8CE6DB40>", "set_training_mode": "<function TD3Policy.set_training_mode at 0x0000020A8CE6DBD0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x0000020A8CE71E80>"}, "verbose": 5, "policy_kwargs": {"net_arch": [400, 300]}, "num_timesteps": 1000227, "_total_timesteps": 1000000.0, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1704195015053101600, "learning_rate": 0.001, "tensorboard_log": "runs/BipedalWalker-v3__td3-BipedalWalker-v3__1__1704194999", "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWV1QAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJZgAAAAAAAAABxRfr4Uvt08kDHEv7ZCqL4NTzU+2FhAvKKtOUC5Syw+HPWGP07yZcDcQDe91m5YP+RK87zZt7Q/t4zGP1O0xj+XgsY/CAHGPxDuxD8CF8M/pCPAP53Suz8eayQ/1OYWPpSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsYhpSMAUOUdJRSlC4="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdAAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYBAAAAAAAAAAGUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwGFlIwBQ5R0lFKULg=="}, "_last_original_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWV1QAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJZgAAAAAAAAAJz3Mzt3Bhk3ZeKbuo4Sg7xcVL098u80OzYeXD+dt0+6AACAPzE9Bj167jQ7noJaP0c59roAAIA/S7LhPnRC5D6FP+w+O6b6Pum6CD/VOho/t4o1P1jMYj8AAIA/AACAP5SMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsYhpSMAUOUdJRSlC4="}, "_episode_num": 1289, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.00022699999999997722, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHMbzO5avA6MAWyUTbcCjAF0lEdA0AV65E+gUXV9lChoBkdAcyHQ79ycTmgHTa4CaAhHQNAI2M/yGzt1fZQoaAZHQHMImdqcmShoB03JAmgIR0DQC7Ul4TsZdX2UKGgGR0BzCfCtRvWIaAdNvwJoCEdA0A66v1lGw3V9lChoBkdAcyYWXkYGdWgHTakCaAhHQNARlLL2YfJ1fZQoaAZHQHMGzrmhdt5oB020AmgIR0DQFEehnJ1adX2UKGgGR0BzAqdAgPmQaAdNyAJoCEdA0BcLgV45cXV9lChoBkdAcxkqSX+l02gHTaYCaAhHQNAZ4IPoV211fZQoaAZHQHMLobfgrH5oB02zAmgIR0DQHIpvegtfdX2UKGgGR0BzDXU4JeE7aAdNmgJoCEdA0B9+IDHOr3V9lChoBkdAcxnwXIlt0mgHTbYCaAhHQNAi7qyWzGB1fZQoaAZHQHL99Sde6ZpoB02yAmgIR0DQJkByKekIdX2UKGgGR0BzDCfJ3gUDaAdNmwJoCEdA0Cl88OCoTHV9lChoBkdAcysGorFwUGgHTY4CaAhHQNAsg84cWCV1fZQoaAZHQHL7fqTr3TNoB02xAmgIR0DQL6TO7g89dX2UKGgGR0BzA16zE74jaAdNsgJoCEdA0DOKb1RLsnV9lChoBkdAcwwXXyy2QWgHTagCaAhHQNA2hvbsWwh1fZQoaAZHQHMWLvPTodNoB02rAmgIR0DQOaUJx//edX2UKGgGR0BzFQZYPoV3aAdNqQJoCEdA0DzIC+lCTnV9lChoBkdAcxnu14Pf9GgHTaICaAhHQNBARVEAo5R1fZQoaAZHQHMT238XN1RoB02hAmgIR0DQRAF4+r2hdX2UKGgGR0BzBIiY9gWraAdNugJoCEdA0EcrOCGvfXV9lChoBkdAcwXRgJC0GGgHTaoCaAhHQNBKberyUcJ1fZQoaAZHQHMGRU3n6mBoB023AmgIR0DQTaxJg9eQdX2UKGgGR0BzHYDW9US7aAdNqQJoCEdA0FEUweeWfXV9lChoBkdAcuuAkcCHRGgHTcsCaAhHQNBUfymygPF1fZQoaAZHQHMINcfNiYtoB02vAmgIR0DQV8svDgqFdX2UKGgGR0BzKWbPQfITaAdNqgJoCEdA0FscMoc7yXV9lChoBkdAcvZ/Nqxkd2gHTa8CaAhHQNBeRTmGM4t1fZQoaAZHQHLbgL3K0UpoB026AmgIR0DQYWCqlxffdX2UKGgGR0BzGGXBxgiNaAdNsgJoCEdA0GTLUh3aBnV9lChoBkdAcutxO+IuXmgHTbcCaAhHQNBoDiV0Lc91fZQoaAZHQHMh7+T/yXloB022AmgIR0DQazxLRKHxdX2UKGgGR0BzL1kZrHlwaAdNpgJoCEdA0G5xljVhC3V9lChoBkdAcytYv38GcGgHTa0CaAhHQNBxlDVpbll1fZQoaAZHQHMue4gA6uJoB024AmgIR0DQdNhoYekpdX2UKGgGR0BzGA2aUiY+aAdNxAJoCEdA0Hg7aaTfSHV9lChoBkdAcwvHKwIMSmgHTbMCaAhHQNB7lNLL6k91fZQoaAZHQHMce9Jz1btoB020AmgIR0DQfs0xWT5gdX2UKGgGR0BzG75xiobXaAdNpwJoCEdA0IIwj9XLeXV9lChoBkdAcyIWNFSbY2gHTcQCaAhHQNCFXgaJhv11fZQoaAZHQHMrYsunMt9oB02oAmgIR0DQiKQspXp4dX2UKGgGR0Byv2J0nw5OaAdNzQJoCEdA0IvFzoEB83V9lChoBkdAcwc5Ec81XWgHTaoCaAhHQNCPFYChew91fZQoaAZHQHMNEpd8iOhoB023AmgIR0DQkk2HdoFndX2UKGgGR0BzF1sdkrf+aAdNqwJoCEdA0JV83A2ycHV9lChoBkdAcxZe2d/ax2gHTaYCaAhHQNCYm9OuaF51fZQoaAZHQHMgK9kBjnVoB02wAmgIR0DQm84XQ+lkdX2UKGgGR0BX2IRZlnRLaAdN0QFoCEdA0J76G8mKInV9lChoBkdAcxRYwIt16mgHTaQCaAhHQNChLhVIZqF1fZQoaAZHQHNEYa5wwTNoB02TAmgIR0DQpEu2RaHLdX2UKGgGR0BzSxd3Sro4aAdNmAJoCEdA0KddO/tY0XV9lChoBkdAcwZ+W4Vh1GgHTaUCaAhHQNCqbVgc94h1fZQoaAZHQHMXBoVVPvdoB02mAmgIR0DQrZ3yH2ytdX2UKGgGR0BzFllvqC6IaAdNtQJoCEdA0LDK51vETHV9lChoBkdAcwBrVvuPWGgHTbICaAhHQNC0NsUuctp1fZQoaAZHQHNB/xDst05oB02lAmgIR0DQt1jf642CdX2UKGgGR0A8FeBxxT86aAdNbwFoCEdA0LpqvjOs1nV9lChoBkdAcza+5vtMPGgHTZkCaAhHQNC8OYE8q4J1fZQoaAZHQHMEbedkJ8hoB025AmgIR0DQv1YgA6uGdX2UKGgGR0BzRkG5c1O1aAdNogJoCEdA0MKTZ7ojfXV9lChoBkdAcutvK2a2F2gHTc0CaAhHQNDF36iGnGd1fZQoaAZHQHMtK0hNdqtoB02jAmgIR0DQyTwTrVvudX2UKGgGR0BzHZGoaUA1aAdNuwJoCEdA0MxfTKDCg3V9lChoBkdAcuDv0yxiX2gHTcsCaAhHQNDPnehwl0J1fZQoaAZHQHMY3z+WGAVoB029AmgIR0DQ0wqNdZ7pdX2UKGgGR0BzI7DuSfUXaAdNsgJoCEdA0NYj3S8aoHV9lChoBkdAczqkwvg3tWgHTaACaAhHQNDZSIcBEKF1fZQoaAZHQHMzOEZiuuBoB02vAmgIR0DQ3E9oh6jWdX2UKGgGR0BzJDuuzQeFaAdNuAJoCEdA0N9uhW5panV9lChoBkdAcx8lhPTG52gHTa8CaAhHQNDioXFPznR1fZQoaAZHQHNBhYzSCvpoB02qAmgIR0DQ5c9HavicdX2UKGgGR0BzRCLVFx4qaAdNpgJoCEdA0OkQaQ3gk3V9lChoBkdAc0QAgPmPo2gHTaoCaAhHQNDsJrqD9O11fZQoaAZHQHMoI/iYLLJoB025AmgIR0DQ74TAxi5NdX2UKGgGR0BzBYAT7EYPaAdN1AJoCEdA0PLnpV0cO3V9lChoBkdAczUTmW+oL2gHTbUCaAhHQND2W6TKT0R1fZQoaAZHQHM7E+HJtBRoB02tAmgIR0DQ+aSSNfgKdX2UKGgGR0BzE0bFS88LaAdNrQJoCEdA0PzPqEOAiHV9lChoBkdAc0LiUxEfDGgHTaICaAhHQNEAGGUSqVB1fZQoaAZHQHMcoGlhw2loB03AAmgIR0DRAyqe8PFvdX2UKGgGR0BzK3cafjCIaAdNvQJoCEdA0QZ3svZh8nV9lChoBkdAcxpC9h7VrmgHTbgCaAhHQNEJqujM3ZR1fZQoaAZHQHNKDtXxOL1oB02lAmgIR0DRDPh7eEZjdX2UKGgGR0BzOGB3A2ycaAdNmwJoCEdA0RARhWYF7nV9lChoBkdAczAbGWD6FmgHTbICaAhHQNETE0p7TlV1fZQoaAZHQHNM33xnWatoB02PAmgIR0DRFkkZ2pyZdX2UKGgGR0BzH83WFvhqaAdNwgJoCEdA0Rnnb0e2eHV9lChoBkdAcz4dU83dbmgHTacCaAhHQNEdciyIHkd1fZQoaAZHQHL0cY2sJY1oB03XAmgIR0DRINvkhib2dX2UKGgGR0BzRowYcebNaAdNswJoCEdA0SRpBDG96HV9lChoBkdAc0duwosqa2gHTakCaAhHQNEnxumJm/Z1fZQoaAZHQHMa18b70nRoB03LAmgIR0DRKxa1eBxxdX2UKGgGR0BzJDZSNwR5aAdNsgJoCEdA0S55RuCPIXV9lChoBkdAcyE0Lc9GJGgHTbwCaAhHQNExsK6asp51fZQoaAZHQHL0WPT5O8FoB03BAmgIR0DRNRKhysCDdX2UKGgGR0BzMh+RYA80aAdNqgJoCEdA0ThnFUQ043V9lChoBkdAc0Gnied072gHTZgCaAhHQNE7prnDBM11fZQoaAZHQHMmohpxm05oB02rAmgIR0DRPsvWrfcfdX2UKGgGR0BzSFv1lGwzaAdNmwJoCEdA0UIDwEhaDHVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 991542, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVTAQAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWGAAAAAAAAAABAQEBAQEBAQEBAQEBAQEBAQEBAQEBAQGUaAiMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLGIWUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBEolhgAAAAAAAAAAQEBAQEBAQEBAQEBAQEBAQEBAQEBAQEBlGgVSxiFlGgZdJRSlIwGX3NoYXBllEsYhZSMA2xvd5RoESiWYAAAAAAAAADbD0nAAACgwAAAoMAAAKDA2w9JwAAAoMDbD0nAAACgwAAAAIDbD0nAAACgwNsPScAAAKDAAAAAgAAAgL8AAIC/AACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAvwAAgL+UaAtLGIWUaBl0lFKUjARoaWdolGgRKJZgAAAAAAAAANsPSUAAAKBAAACgQAAAoEDbD0lAAACgQNsPSUAAAKBAAACgQNsPSUAAAKBA2w9JQAAAoEAAAKBAAACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAP5RoC0sYhZRoGXSUUpSMCGxvd19yZXBylIz+Wy0zLjE0MTU5MjcgLTUuICAgICAgICAtNS4gICAgICAgIC01LiAgICAgICAgLTMuMTQxNTkyNyAtNS4KIC0zLjE0MTU5MjcgLTUuICAgICAgICAtMC4gICAgICAgIC0zLjE0MTU5MjcgLTUuICAgICAgICAtMy4xNDE1OTI3CiAtNS4gICAgICAgIC0wLiAgICAgICAgLTEuICAgICAgICAtMS4gICAgICAgIC0xLiAgICAgICAgLTEuCiAtMS4gICAgICAgIC0xLiAgICAgICAgLTEuICAgICAgICAtMS4gICAgICAgIC0xLiAgICAgICAgLTEuICAgICAgIF2UjAloaWdoX3JlcHKUjOZbMy4xNDE1OTI3IDUuICAgICAgICA1LiAgICAgICAgNS4gICAgICAgIDMuMTQxNTkyNyA1LiAgICAgICAgMy4xNDE1OTI3CiA1LiAgICAgICAgNS4gICAgICAgIDMuMTQxNTkyNyA1LiAgICAgICAgMy4xNDE1OTI3IDUuICAgICAgICA1LgogMS4gICAgICAgIDEuICAgICAgICAxLiAgICAgICAgMS4gICAgICAgIDEuICAgICAgICAxLiAgICAgICAgMS4KIDEuICAgICAgICAxLiAgICAgICAgMS4gICAgICAgXZSMCl9ucF9yYW5kb22UTnViLg==", "dtype": "float32", "bounded_below": "[ True True True True True True True True True True True True\n True True True True True True True True True True True True]", "bounded_above": "[ True True True True True True True True True True True True\n True True True True True True True True True True True True]", "_shape": [24], "low": "[-3.1415927 -5. -5. -5. -3.1415927 -5.\n -3.1415927 -5. -0. -3.1415927 -5. -3.1415927\n -5. -0. -1. -1. -1. -1.\n -1. -1. -1. -1. -1. -1. ]", "high": "[3.1415927 5. 5. 5. 3.1415927 5. 3.1415927\n 5. 5. 3.1415927 5. 3.1415927 5. 5.\n 1. 1. 1. 1. 1. 1. 1.\n 1. 1. 1. ]", "low_repr": "[-3.1415927 -5. -5. -5. -3.1415927 -5.\n -3.1415927 -5. -0. -3.1415927 -5. -3.1415927\n -5. -0. -1. -1. -1. -1.\n -1. -1. -1. -1. -1. -1. ]", "high_repr": "[3.1415927 5. 5. 5. 3.1415927 5. 3.1415927\n 5. 5. 3.1415927 5. 3.1415927 5. 5.\n 1. 1. 1. 1. 1. 1. 1.\n 1. 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVawIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWBAAAAAAAAAABAQEBlGgIjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwSFlIwBQ5R0lFKUjA1ib3VuZGVkX2Fib3ZllGgRKJYEAAAAAAAAAAEBAQGUaBVLBIWUaBl0lFKUjAZfc2hhcGWUSwSFlIwDbG93lGgRKJYQAAAAAAAAAAAAgL8AAIC/AACAvwAAgL+UaAtLBIWUaBl0lFKUjARoaWdolGgRKJYQAAAAAAAAAAAAgD8AAIA/AACAPwAAgD+UaAtLBIWUaBl0lFKUjAhsb3dfcmVwcpSMBC0xLjCUjAloaWdoX3JlcHKUjAMxLjCUjApfbnBfcmFuZG9tlIwUbnVtcHkucmFuZG9tLl9waWNrbGWUjBBfX2dlbmVyYXRvcl9jdG9ylJOUjAVQQ0c2NJRoMowUX19iaXRfZ2VuZXJhdG9yX2N0b3KUk5SGlFKUfZQojA1iaXRfZ2VuZXJhdG9ylIwFUENHNjSUjAVzdGF0ZZR9lChoPYoQJY3Zi+ogH88AozHI4lO5EowDaW5jlIoRn9i2fAYPb6S5z1P3M1g1gAB1jApoYXNfdWludDMylEsAjAh1aW50ZWdlcpRLAHVidWIu", "dtype": "float32", "bounded_below": "[ True True True True]", "bounded_above": "[ True True True True]", "_shape": [4], "low": "[-1. -1. -1. -1.]", "high": "[1. 1. 1. 1.]", "low_repr": "-1.0", "high_repr": "1.0", "_np_random": "Generator(PCG64)"}, "n_envs": 1, "buffer_size": 200000, "batch_size": 100, "learning_starts": 10000, "tau": 0.005, "gamma": 0.98, "gradient_steps": -1, "optimize_memory_usage": false, "replay_buffer_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVNQAAAAAAAACMIHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5idWZmZXJzlIwMUmVwbGF5QnVmZmVylJOULg==", "__module__": "stable_baselines3.common.buffers", "__annotations__": "{'observations': <class 'numpy.ndarray'>, 'next_observations': <class 'numpy.ndarray'>, 'actions': <class 'numpy.ndarray'>, 'rewards': <class 'numpy.ndarray'>, 'dones': <class 'numpy.ndarray'>, 'timeouts': <class 'numpy.ndarray'>}", "__doc__": "\n Replay buffer used in off-policy algorithms like SAC/TD3.\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n of the replay buffer which reduces by almost a factor two the memory used,\n at a cost of more complexity.\n See https://github.com/DLR-RM/stable-baselines3/issues/37#issuecomment-637501195\n and https://github.com/DLR-RM/stable-baselines3/pull/28#issuecomment-637559274\n Cannot be used in combination with handle_timeout_termination.\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n ", "__init__": "<function ReplayBuffer.__init__ at 0x0000020A8CB9C4C0>", "add": "<function ReplayBuffer.add at 0x0000020A8CB9C550>", "sample": "<function ReplayBuffer.sample at 0x0000020A8CB9C5E0>", "_get_samples": "<function ReplayBuffer._get_samples at 0x0000020A8CB9C670>", "_maybe_cast_dtype": "<staticmethod(<function ReplayBuffer._maybe_cast_dtype at 0x0000020A8CB9C700>)>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x0000020A8CB99940>"}, "replay_buffer_kwargs": {}, "train_freq": {":type:": "<class 'stable_baselines3.common.type_aliases.TrainFreq'>", ":serialized:": "gAWVZAAAAAAAAACMJXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi50eXBlX2FsaWFzZXOUjAlUcmFpbkZyZXGUk5RLAWgAjBJUcmFpbkZyZXF1ZW5jeVVuaXSUk5SMB2VwaXNvZGWUhZRSlIaUgZQu"}, "use_sde_at_warmup": false, "policy_delay": 2, "target_noise_clip": 0.5, "target_policy_noise": 0.2, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVaAIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMU2M6XFVzZXJzXHByaWNvXC5jb25kYVxlbnZzXEFUTk5cbGliXHNpdGUtcGFja2FnZXNcc3RhYmxlX2Jhc2VsaW5lczNcY29tbW9uXHV0aWxzLnB5lIwEZnVuY5RLg0MCBAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB59lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz9QYk3S8an8hZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "actor_batch_norm_stats": [], "critic_batch_norm_stats": [], "actor_batch_norm_stats_target": [], "critic_batch_norm_stats_target": [], "system_info": {"OS": "Windows-10-10.0.22621-SP0 10.0.22621", "Python": "3.10.12", "Stable-Baselines3": "2.2.1", "PyTorch": "2.1.0", "GPU Enabled": "False", "Numpy": "1.26.0", "Cloudpickle": "3.0.0", "Gymnasium": "0.28.1", "OpenAI Gym": "0.21.0"}}
logs/TD3_1/events.out.tfevents.1704195015.Thunder.12240.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad575e7878ede9a12fb05d32f8b27b9778aac5575e2ee4ad4034251925ac215b
3
+ size 108375
logs/events.out.tfevents.1704195015.Thunder.12240.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c54897fbabee690214adcd1ffb7a0a866c08075cb0f0ac63ead93b4d37137e0c
3
+ size 441
replay.mp4 ADDED
Binary file (356 kB). View file
 
results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"mean_reward": 306.7897345, "std_reward": 1.199929253409739, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-01-02T18:24:55.601456"}
td3-BipedalWalker-v3.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4868daab7b23f0bb433c9fde4ce917d0288a01be4ea2c7ed4d23a40058acbf5c
3
+ size 6379605
td3-BipedalWalker-v3/_stable_baselines3_version ADDED
@@ -0,0 +1 @@
 
 
1
+ 2.2.1
td3-BipedalWalker-v3/actor.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2da27a0541b2958dd2629565110429db4f65423b453ca0bbcfaef8eec50e37d
3
+ size 1057184
td3-BipedalWalker-v3/critic.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e9f3acd35b44f5d4f04a91ec0b20bd137a243c1142148b8e02bc159efe639ca
3
+ size 2124522
td3-BipedalWalker-v3/data ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "policy_class": {
3
+ ":type:": "<class 'abc.ABCMeta'>",
4
+ ":serialized:": "gAWVMAAAAAAAAACMHnN0YWJsZV9iYXNlbGluZXMzLnRkMy5wb2xpY2llc5SMCVREM1BvbGljeZSTlC4=",
5
+ "__module__": "stable_baselines3.td3.policies",
6
+ "__annotations__": "{'actor': <class 'stable_baselines3.td3.policies.Actor'>, 'actor_target': <class 'stable_baselines3.td3.policies.Actor'>, 'critic': <class 'stable_baselines3.common.policies.ContinuousCritic'>, 'critic_target': <class 'stable_baselines3.common.policies.ContinuousCritic'>}",
7
+ "__doc__": "\n Policy class (with both actor and critic) for TD3.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ",
8
+ "__init__": "<function TD3Policy.__init__ at 0x0000020A8CE6D7E0>",
9
+ "_build": "<function TD3Policy._build at 0x0000020A8CE6D870>",
10
+ "_get_constructor_parameters": "<function TD3Policy._get_constructor_parameters at 0x0000020A8CE6D900>",
11
+ "make_actor": "<function TD3Policy.make_actor at 0x0000020A8CE6D990>",
12
+ "make_critic": "<function TD3Policy.make_critic at 0x0000020A8CE6DA20>",
13
+ "forward": "<function TD3Policy.forward at 0x0000020A8CE6DAB0>",
14
+ "_predict": "<function TD3Policy._predict at 0x0000020A8CE6DB40>",
15
+ "set_training_mode": "<function TD3Policy.set_training_mode at 0x0000020A8CE6DBD0>",
16
+ "__abstractmethods__": "frozenset()",
17
+ "_abc_impl": "<_abc._abc_data object at 0x0000020A8CE71E80>"
18
+ },
19
+ "verbose": 5,
20
+ "policy_kwargs": {
21
+ "net_arch": [
22
+ 400,
23
+ 300
24
+ ]
25
+ },
26
+ "num_timesteps": 1000227,
27
+ "_total_timesteps": 1000000.0,
28
+ "_num_timesteps_at_start": 0,
29
+ "seed": null,
30
+ "action_noise": null,
31
+ "start_time": 1704195015053101600,
32
+ "learning_rate": 0.001,
33
+ "tensorboard_log": "runs/BipedalWalker-v3__td3-BipedalWalker-v3__1__1704194999",
34
+ "_last_obs": {
35
+ ":type:": "<class 'numpy.ndarray'>",
36
+ ":serialized:": "gAWV1QAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJZgAAAAAAAAABxRfr4Uvt08kDHEv7ZCqL4NTzU+2FhAvKKtOUC5Syw+HPWGP07yZcDcQDe91m5YP+RK87zZt7Q/t4zGP1O0xj+XgsY/CAHGPxDuxD8CF8M/pCPAP53Suz8eayQ/1OYWPpSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsYhpSMAUOUdJRSlC4="
37
+ },
38
+ "_last_episode_starts": {
39
+ ":type:": "<class 'numpy.ndarray'>",
40
+ ":serialized:": "gAWVdAAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYBAAAAAAAAAAGUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwGFlIwBQ5R0lFKULg=="
41
+ },
42
+ "_last_original_obs": {
43
+ ":type:": "<class 'numpy.ndarray'>",
44
+ ":serialized:": "gAWV1QAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJZgAAAAAAAAAJz3Mzt3Bhk3ZeKbuo4Sg7xcVL098u80OzYeXD+dt0+6AACAPzE9Bj167jQ7noJaP0c59roAAIA/S7LhPnRC5D6FP+w+O6b6Pum6CD/VOho/t4o1P1jMYj8AAIA/AACAP5SMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsYhpSMAUOUdJRSlC4="
45
+ },
46
+ "_episode_num": 1289,
47
+ "use_sde": false,
48
+ "sde_sample_freq": -1,
49
+ "_current_progress_remaining": -0.00022699999999997722,
50
+ "_stats_window_size": 100,
51
+ "ep_info_buffer": {
52
+ ":type:": "<class 'collections.deque'>",
53
+ ":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHMbzO5avA6MAWyUTbcCjAF0lEdA0AV65E+gUXV9lChoBkdAcyHQ79ycTmgHTa4CaAhHQNAI2M/yGzt1fZQoaAZHQHMImdqcmShoB03JAmgIR0DQC7Ul4TsZdX2UKGgGR0BzCfCtRvWIaAdNvwJoCEdA0A66v1lGw3V9lChoBkdAcyYWXkYGdWgHTakCaAhHQNARlLL2YfJ1fZQoaAZHQHMGzrmhdt5oB020AmgIR0DQFEehnJ1adX2UKGgGR0BzAqdAgPmQaAdNyAJoCEdA0BcLgV45cXV9lChoBkdAcxkqSX+l02gHTaYCaAhHQNAZ4IPoV211fZQoaAZHQHMLobfgrH5oB02zAmgIR0DQHIpvegtfdX2UKGgGR0BzDXU4JeE7aAdNmgJoCEdA0B9+IDHOr3V9lChoBkdAcxnwXIlt0mgHTbYCaAhHQNAi7qyWzGB1fZQoaAZHQHL99Sde6ZpoB02yAmgIR0DQJkByKekIdX2UKGgGR0BzDCfJ3gUDaAdNmwJoCEdA0Cl88OCoTHV9lChoBkdAcysGorFwUGgHTY4CaAhHQNAsg84cWCV1fZQoaAZHQHL7fqTr3TNoB02xAmgIR0DQL6TO7g89dX2UKGgGR0BzA16zE74jaAdNsgJoCEdA0DOKb1RLsnV9lChoBkdAcwwXXyy2QWgHTagCaAhHQNA2hvbsWwh1fZQoaAZHQHMWLvPTodNoB02rAmgIR0DQOaUJx//edX2UKGgGR0BzFQZYPoV3aAdNqQJoCEdA0DzIC+lCTnV9lChoBkdAcxnu14Pf9GgHTaICaAhHQNBARVEAo5R1fZQoaAZHQHMT238XN1RoB02hAmgIR0DQRAF4+r2hdX2UKGgGR0BzBIiY9gWraAdNugJoCEdA0EcrOCGvfXV9lChoBkdAcwXRgJC0GGgHTaoCaAhHQNBKberyUcJ1fZQoaAZHQHMGRU3n6mBoB023AmgIR0DQTaxJg9eQdX2UKGgGR0BzHYDW9US7aAdNqQJoCEdA0FEUweeWfXV9lChoBkdAcuuAkcCHRGgHTcsCaAhHQNBUfymygPF1fZQoaAZHQHMINcfNiYtoB02vAmgIR0DQV8svDgqFdX2UKGgGR0BzKWbPQfITaAdNqgJoCEdA0FscMoc7yXV9lChoBkdAcvZ/Nqxkd2gHTa8CaAhHQNBeRTmGM4t1fZQoaAZHQHLbgL3K0UpoB026AmgIR0DQYWCqlxffdX2UKGgGR0BzGGXBxgiNaAdNsgJoCEdA0GTLUh3aBnV9lChoBkdAcutxO+IuXmgHTbcCaAhHQNBoDiV0Lc91fZQoaAZHQHMh7+T/yXloB022AmgIR0DQazxLRKHxdX2UKGgGR0BzL1kZrHlwaAdNpgJoCEdA0G5xljVhC3V9lChoBkdAcytYv38GcGgHTa0CaAhHQNBxlDVpbll1fZQoaAZHQHMue4gA6uJoB024AmgIR0DQdNhoYekpdX2UKGgGR0BzGA2aUiY+aAdNxAJoCEdA0Hg7aaTfSHV9lChoBkdAcwvHKwIMSmgHTbMCaAhHQNB7lNLL6k91fZQoaAZHQHMce9Jz1btoB020AmgIR0DQfs0xWT5gdX2UKGgGR0BzG75xiobXaAdNpwJoCEdA0IIwj9XLeXV9lChoBkdAcyIWNFSbY2gHTcQCaAhHQNCFXgaJhv11fZQoaAZHQHMrYsunMt9oB02oAmgIR0DQiKQspXp4dX2UKGgGR0Byv2J0nw5OaAdNzQJoCEdA0IvFzoEB83V9lChoBkdAcwc5Ec81XWgHTaoCaAhHQNCPFYChew91fZQoaAZHQHMNEpd8iOhoB023AmgIR0DQkk2HdoFndX2UKGgGR0BzF1sdkrf+aAdNqwJoCEdA0JV83A2ycHV9lChoBkdAcxZe2d/ax2gHTaYCaAhHQNCYm9OuaF51fZQoaAZHQHMgK9kBjnVoB02wAmgIR0DQm84XQ+lkdX2UKGgGR0BX2IRZlnRLaAdN0QFoCEdA0J76G8mKInV9lChoBkdAcxRYwIt16mgHTaQCaAhHQNChLhVIZqF1fZQoaAZHQHNEYa5wwTNoB02TAmgIR0DQpEu2RaHLdX2UKGgGR0BzSxd3Sro4aAdNmAJoCEdA0KddO/tY0XV9lChoBkdAcwZ+W4Vh1GgHTaUCaAhHQNCqbVgc94h1fZQoaAZHQHMXBoVVPvdoB02mAmgIR0DQrZ3yH2ytdX2UKGgGR0BzFllvqC6IaAdNtQJoCEdA0LDK51vETHV9lChoBkdAcwBrVvuPWGgHTbICaAhHQNC0NsUuctp1fZQoaAZHQHNB/xDst05oB02lAmgIR0DQt1jf642CdX2UKGgGR0A8FeBxxT86aAdNbwFoCEdA0LpqvjOs1nV9lChoBkdAcza+5vtMPGgHTZkCaAhHQNC8OYE8q4J1fZQoaAZHQHMEbedkJ8hoB025AmgIR0DQv1YgA6uGdX2UKGgGR0BzRkG5c1O1aAdNogJoCEdA0MKTZ7ojfXV9lChoBkdAcutvK2a2F2gHTc0CaAhHQNDF36iGnGd1fZQoaAZHQHMtK0hNdqtoB02jAmgIR0DQyTwTrVvudX2UKGgGR0BzHZGoaUA1aAdNuwJoCEdA0MxfTKDCg3V9lChoBkdAcuDv0yxiX2gHTcsCaAhHQNDPnehwl0J1fZQoaAZHQHMY3z+WGAVoB029AmgIR0DQ0wqNdZ7pdX2UKGgGR0BzI7DuSfUXaAdNsgJoCEdA0NYj3S8aoHV9lChoBkdAczqkwvg3tWgHTaACaAhHQNDZSIcBEKF1fZQoaAZHQHMzOEZiuuBoB02vAmgIR0DQ3E9oh6jWdX2UKGgGR0BzJDuuzQeFaAdNuAJoCEdA0N9uhW5panV9lChoBkdAcx8lhPTG52gHTa8CaAhHQNDioXFPznR1fZQoaAZHQHNBhYzSCvpoB02qAmgIR0DQ5c9HavicdX2UKGgGR0BzRCLVFx4qaAdNpgJoCEdA0OkQaQ3gk3V9lChoBkdAc0QAgPmPo2gHTaoCaAhHQNDsJrqD9O11fZQoaAZHQHMoI/iYLLJoB025AmgIR0DQ74TAxi5NdX2UKGgGR0BzBYAT7EYPaAdN1AJoCEdA0PLnpV0cO3V9lChoBkdAczUTmW+oL2gHTbUCaAhHQND2W6TKT0R1fZQoaAZHQHM7E+HJtBRoB02tAmgIR0DQ+aSSNfgKdX2UKGgGR0BzE0bFS88LaAdNrQJoCEdA0PzPqEOAiHV9lChoBkdAc0LiUxEfDGgHTaICaAhHQNEAGGUSqVB1fZQoaAZHQHMcoGlhw2loB03AAmgIR0DRAyqe8PFvdX2UKGgGR0BzK3cafjCIaAdNvQJoCEdA0QZ3svZh8nV9lChoBkdAcxpC9h7VrmgHTbgCaAhHQNEJqujM3ZR1fZQoaAZHQHNKDtXxOL1oB02lAmgIR0DRDPh7eEZjdX2UKGgGR0BzOGB3A2ycaAdNmwJoCEdA0RARhWYF7nV9lChoBkdAczAbGWD6FmgHTbICaAhHQNETE0p7TlV1fZQoaAZHQHNM33xnWatoB02PAmgIR0DRFkkZ2pyZdX2UKGgGR0BzH83WFvhqaAdNwgJoCEdA0Rnnb0e2eHV9lChoBkdAcz4dU83dbmgHTacCaAhHQNEdciyIHkd1fZQoaAZHQHL0cY2sJY1oB03XAmgIR0DRINvkhib2dX2UKGgGR0BzRowYcebNaAdNswJoCEdA0SRpBDG96HV9lChoBkdAc0duwosqa2gHTakCaAhHQNEnxumJm/Z1fZQoaAZHQHMa18b70nRoB03LAmgIR0DRKxa1eBxxdX2UKGgGR0BzJDZSNwR5aAdNsgJoCEdA0S55RuCPIXV9lChoBkdAcyE0Lc9GJGgHTbwCaAhHQNExsK6asp51fZQoaAZHQHL0WPT5O8FoB03BAmgIR0DRNRKhysCDdX2UKGgGR0BzMh+RYA80aAdNqgJoCEdA0ThnFUQ043V9lChoBkdAc0Gnied072gHTZgCaAhHQNE7prnDBM11fZQoaAZHQHMmohpxm05oB02rAmgIR0DRPsvWrfcfdX2UKGgGR0BzSFv1lGwzaAdNmwJoCEdA0UIDwEhaDHVlLg=="
54
+ },
55
+ "ep_success_buffer": {
56
+ ":type:": "<class 'collections.deque'>",
57
+ ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
58
+ },
59
+ "_n_updates": 991542,
60
+ "observation_space": {
61
+ ":type:": "<class 'gymnasium.spaces.box.Box'>",
62
+ ":serialized:": "gAWVTAQAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWGAAAAAAAAAABAQEBAQEBAQEBAQEBAQEBAQEBAQEBAQGUaAiMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLGIWUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBEolhgAAAAAAAAAAQEBAQEBAQEBAQEBAQEBAQEBAQEBAQEBlGgVSxiFlGgZdJRSlIwGX3NoYXBllEsYhZSMA2xvd5RoESiWYAAAAAAAAADbD0nAAACgwAAAoMAAAKDA2w9JwAAAoMDbD0nAAACgwAAAAIDbD0nAAACgwNsPScAAAKDAAAAAgAAAgL8AAIC/AACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAvwAAgL+UaAtLGIWUaBl0lFKUjARoaWdolGgRKJZgAAAAAAAAANsPSUAAAKBAAACgQAAAoEDbD0lAAACgQNsPSUAAAKBAAACgQNsPSUAAAKBA2w9JQAAAoEAAAKBAAACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/AACAP5RoC0sYhZRoGXSUUpSMCGxvd19yZXBylIz+Wy0zLjE0MTU5MjcgLTUuICAgICAgICAtNS4gICAgICAgIC01LiAgICAgICAgLTMuMTQxNTkyNyAtNS4KIC0zLjE0MTU5MjcgLTUuICAgICAgICAtMC4gICAgICAgIC0zLjE0MTU5MjcgLTUuICAgICAgICAtMy4xNDE1OTI3CiAtNS4gICAgICAgIC0wLiAgICAgICAgLTEuICAgICAgICAtMS4gICAgICAgIC0xLiAgICAgICAgLTEuCiAtMS4gICAgICAgIC0xLiAgICAgICAgLTEuICAgICAgICAtMS4gICAgICAgIC0xLiAgICAgICAgLTEuICAgICAgIF2UjAloaWdoX3JlcHKUjOZbMy4xNDE1OTI3IDUuICAgICAgICA1LiAgICAgICAgNS4gICAgICAgIDMuMTQxNTkyNyA1LiAgICAgICAgMy4xNDE1OTI3CiA1LiAgICAgICAgNS4gICAgICAgIDMuMTQxNTkyNyA1LiAgICAgICAgMy4xNDE1OTI3IDUuICAgICAgICA1LgogMS4gICAgICAgIDEuICAgICAgICAxLiAgICAgICAgMS4gICAgICAgIDEuICAgICAgICAxLiAgICAgICAgMS4KIDEuICAgICAgICAxLiAgICAgICAgMS4gICAgICAgXZSMCl9ucF9yYW5kb22UTnViLg==",
63
+ "dtype": "float32",
64
+ "bounded_below": "[ True True True True True True True True True True True True\n True True True True True True True True True True True True]",
65
+ "bounded_above": "[ True True True True True True True True True True True True\n True True True True True True True True True True True True]",
66
+ "_shape": [
67
+ 24
68
+ ],
69
+ "low": "[-3.1415927 -5. -5. -5. -3.1415927 -5.\n -3.1415927 -5. -0. -3.1415927 -5. -3.1415927\n -5. -0. -1. -1. -1. -1.\n -1. -1. -1. -1. -1. -1. ]",
70
+ "high": "[3.1415927 5. 5. 5. 3.1415927 5. 3.1415927\n 5. 5. 3.1415927 5. 3.1415927 5. 5.\n 1. 1. 1. 1. 1. 1. 1.\n 1. 1. 1. ]",
71
+ "low_repr": "[-3.1415927 -5. -5. -5. -3.1415927 -5.\n -3.1415927 -5. -0. -3.1415927 -5. -3.1415927\n -5. -0. -1. -1. -1. -1.\n -1. -1. -1. -1. -1. -1. ]",
72
+ "high_repr": "[3.1415927 5. 5. 5. 3.1415927 5. 3.1415927\n 5. 5. 3.1415927 5. 3.1415927 5. 5.\n 1. 1. 1. 1. 1. 1. 1.\n 1. 1. 1. ]",
73
+ "_np_random": null
74
+ },
75
+ "action_space": {
76
+ ":type:": "<class 'gymnasium.spaces.box.Box'>",
77
+ ":serialized:": "gAWVawIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWBAAAAAAAAAABAQEBlGgIjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwSFlIwBQ5R0lFKUjA1ib3VuZGVkX2Fib3ZllGgRKJYEAAAAAAAAAAEBAQGUaBVLBIWUaBl0lFKUjAZfc2hhcGWUSwSFlIwDbG93lGgRKJYQAAAAAAAAAAAAgL8AAIC/AACAvwAAgL+UaAtLBIWUaBl0lFKUjARoaWdolGgRKJYQAAAAAAAAAAAAgD8AAIA/AACAPwAAgD+UaAtLBIWUaBl0lFKUjAhsb3dfcmVwcpSMBC0xLjCUjAloaWdoX3JlcHKUjAMxLjCUjApfbnBfcmFuZG9tlIwUbnVtcHkucmFuZG9tLl9waWNrbGWUjBBfX2dlbmVyYXRvcl9jdG9ylJOUjAVQQ0c2NJRoMowUX19iaXRfZ2VuZXJhdG9yX2N0b3KUk5SGlFKUfZQojA1iaXRfZ2VuZXJhdG9ylIwFUENHNjSUjAVzdGF0ZZR9lChoPYoQJY3Zi+ogH88AozHI4lO5EowDaW5jlIoRn9i2fAYPb6S5z1P3M1g1gAB1jApoYXNfdWludDMylEsAjAh1aW50ZWdlcpRLAHVidWIu",
78
+ "dtype": "float32",
79
+ "bounded_below": "[ True True True True]",
80
+ "bounded_above": "[ True True True True]",
81
+ "_shape": [
82
+ 4
83
+ ],
84
+ "low": "[-1. -1. -1. -1.]",
85
+ "high": "[1. 1. 1. 1.]",
86
+ "low_repr": "-1.0",
87
+ "high_repr": "1.0",
88
+ "_np_random": "Generator(PCG64)"
89
+ },
90
+ "n_envs": 1,
91
+ "buffer_size": 200000,
92
+ "batch_size": 100,
93
+ "learning_starts": 10000,
94
+ "tau": 0.005,
95
+ "gamma": 0.98,
96
+ "gradient_steps": -1,
97
+ "optimize_memory_usage": false,
98
+ "replay_buffer_class": {
99
+ ":type:": "<class 'abc.ABCMeta'>",
100
+ ":serialized:": "gAWVNQAAAAAAAACMIHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5idWZmZXJzlIwMUmVwbGF5QnVmZmVylJOULg==",
101
+ "__module__": "stable_baselines3.common.buffers",
102
+ "__annotations__": "{'observations': <class 'numpy.ndarray'>, 'next_observations': <class 'numpy.ndarray'>, 'actions': <class 'numpy.ndarray'>, 'rewards': <class 'numpy.ndarray'>, 'dones': <class 'numpy.ndarray'>, 'timeouts': <class 'numpy.ndarray'>}",
103
+ "__doc__": "\n Replay buffer used in off-policy algorithms like SAC/TD3.\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n of the replay buffer which reduces by almost a factor two the memory used,\n at a cost of more complexity.\n See https://github.com/DLR-RM/stable-baselines3/issues/37#issuecomment-637501195\n and https://github.com/DLR-RM/stable-baselines3/pull/28#issuecomment-637559274\n Cannot be used in combination with handle_timeout_termination.\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n ",
104
+ "__init__": "<function ReplayBuffer.__init__ at 0x0000020A8CB9C4C0>",
105
+ "add": "<function ReplayBuffer.add at 0x0000020A8CB9C550>",
106
+ "sample": "<function ReplayBuffer.sample at 0x0000020A8CB9C5E0>",
107
+ "_get_samples": "<function ReplayBuffer._get_samples at 0x0000020A8CB9C670>",
108
+ "_maybe_cast_dtype": "<staticmethod(<function ReplayBuffer._maybe_cast_dtype at 0x0000020A8CB9C700>)>",
109
+ "__abstractmethods__": "frozenset()",
110
+ "_abc_impl": "<_abc._abc_data object at 0x0000020A8CB99940>"
111
+ },
112
+ "replay_buffer_kwargs": {},
113
+ "train_freq": {
114
+ ":type:": "<class 'stable_baselines3.common.type_aliases.TrainFreq'>",
115
+ ":serialized:": "gAWVZAAAAAAAAACMJXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi50eXBlX2FsaWFzZXOUjAlUcmFpbkZyZXGUk5RLAWgAjBJUcmFpbkZyZXF1ZW5jeVVuaXSUk5SMB2VwaXNvZGWUhZRSlIaUgZQu"
116
+ },
117
+ "use_sde_at_warmup": false,
118
+ "policy_delay": 2,
119
+ "target_noise_clip": 0.5,
120
+ "target_policy_noise": 0.2,
121
+ "lr_schedule": {
122
+ ":type:": "<class 'function'>",
123
+ ":serialized:": "gAWVaAIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMU2M6XFVzZXJzXHByaWNvXC5jb25kYVxlbnZzXEFUTk5cbGliXHNpdGUtcGFja2FnZXNcc3RhYmxlX2Jhc2VsaW5lczNcY29tbW9uXHV0aWxzLnB5lIwEZnVuY5RLg0MCBAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB59lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz9QYk3S8an8hZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"
124
+ },
125
+ "actor_batch_norm_stats": [],
126
+ "critic_batch_norm_stats": [],
127
+ "actor_batch_norm_stats_target": [],
128
+ "critic_batch_norm_stats_target": []
129
+ }
td3-BipedalWalker-v3/policy.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:319a5234a5dccbe38fcedced007c6a319b3de92986fd063b136b28338dae1ca5
3
+ size 3179242
td3-BipedalWalker-v3/pytorch_variables.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb4dde0c1ad63b7740276006a06cc491b21b407ea6c889928c223ec77ddad79f
3
+ size 864
td3-BipedalWalker-v3/system_info.txt ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - OS: Windows-10-10.0.22621-SP0 10.0.22621
2
+ - Python: 3.10.12
3
+ - Stable-Baselines3: 2.2.1
4
+ - PyTorch: 2.1.0
5
+ - GPU Enabled: False
6
+ - Numpy: 1.26.0
7
+ - Cloudpickle: 3.0.0
8
+ - Gymnasium: 0.28.1
9
+ - OpenAI Gym: 0.21.0
vec_normalize.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c5a14fec7ae13a089397d4d3713dca51dd3a1fa3f916f3008913c8d52e21c21
3
+ size 2861