reachrkr commited on
Commit
a6eb7c0
1 Parent(s): d03672d

Initial commit

Browse files
README.md CHANGED
@@ -8,16 +8,17 @@ tags:
8
  model-index:
9
  - name: DQN
10
  results:
11
- - metrics:
12
- - type: mean_reward
13
- value: 533.50 +/- 199.53
14
- name: mean_reward
15
- task:
16
  type: reinforcement-learning
17
  name: reinforcement-learning
18
  dataset:
19
  name: SpaceInvadersNoFrameskip-v4
20
  type: SpaceInvadersNoFrameskip-v4
 
 
 
 
 
21
  ---
22
 
23
  # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4**
@@ -35,17 +36,28 @@ RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/>
35
  SB3: https://github.com/DLR-RM/stable-baselines3<br/>
36
  SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
37
 
 
 
 
 
 
38
  ```
39
  # Download model and save it into the logs/ folder
40
- python -m utils.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga reachrkr -f logs/
41
- python enjoy.py --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
 
 
 
 
 
 
42
  ```
43
 
44
  ## Training (with the RL Zoo)
45
  ```
46
- python train.py --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
47
  # Upload the model and generate video (when possible)
48
- python -m utils.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga reachrkr
49
  ```
50
 
51
  ## Hyperparameters
@@ -60,7 +72,7 @@ OrderedDict([('batch_size', 32),
60
  ('gradient_steps', 1),
61
  ('learning_rate', 0.0001),
62
  ('learning_starts', 100000),
63
- ('n_timesteps', 1000000.0),
64
  ('optimize_memory_usage', False),
65
  ('policy', 'CnnPolicy'),
66
  ('target_update_interval', 1000),
 
8
  model-index:
9
  - name: DQN
10
  results:
11
+ - task:
 
 
 
 
12
  type: reinforcement-learning
13
  name: reinforcement-learning
14
  dataset:
15
  name: SpaceInvadersNoFrameskip-v4
16
  type: SpaceInvadersNoFrameskip-v4
17
+ metrics:
18
+ - type: mean_reward
19
+ value: 210.00 +/- 145.89
20
+ name: mean_reward
21
+ verified: false
22
  ---
23
 
24
  # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4**
 
36
  SB3: https://github.com/DLR-RM/stable-baselines3<br/>
37
  SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
38
 
39
+ Install the RL Zoo (with SB3 and SB3-Contrib):
40
+ ```bash
41
+ pip install rl_zoo3
42
+ ```
43
+
44
  ```
45
  # Download model and save it into the logs/ folder
46
+ python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga reachrkr -f logs/
47
+ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
48
+ ```
49
+
50
+ If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
51
+ ```
52
+ python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga reachrkr -f logs/
53
+ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
54
  ```
55
 
56
  ## Training (with the RL Zoo)
57
  ```
58
+ python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
59
  # Upload the model and generate video (when possible)
60
+ python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga reachrkr
61
  ```
62
 
63
  ## Hyperparameters
 
72
  ('gradient_steps', 1),
73
  ('learning_rate', 0.0001),
74
  ('learning_starts', 100000),
75
+ ('n_timesteps', 100000.0),
76
  ('optimize_memory_usage', False),
77
  ('policy', 'CnnPolicy'),
78
  ('target_update_interval', 1000),
args.yml CHANGED
@@ -1,6 +1,8 @@
1
  !!python/object/apply:collections.OrderedDict
2
  - - - algo
3
  - dqn
 
 
4
  - - device
5
  - auto
6
  - - env
@@ -41,6 +43,8 @@
41
  - null
42
  - - optimize_hyperparameters
43
  - false
 
 
44
  - - pruner
45
  - median
46
  - - sampler
@@ -50,7 +54,7 @@
50
  - - save_replay_buffer
51
  - false
52
  - - seed
53
- - 875814073
54
  - - storage
55
  - null
56
  - - study_name
@@ -73,3 +77,7 @@
73
  - null
74
  - - wandb_project_name
75
  - sb3
 
 
 
 
 
1
  !!python/object/apply:collections.OrderedDict
2
  - - - algo
3
  - dqn
4
+ - - conf_file
5
+ - null
6
  - - device
7
  - auto
8
  - - env
 
43
  - null
44
  - - optimize_hyperparameters
45
  - false
46
+ - - progress
47
+ - false
48
  - - pruner
49
  - median
50
  - - sampler
 
54
  - - save_replay_buffer
55
  - false
56
  - - seed
57
+ - 2794116141
58
  - - storage
59
  - null
60
  - - study_name
 
77
  - null
78
  - - wandb_project_name
79
  - sb3
80
+ - - wandb_tags
81
+ - []
82
+ - - yaml_file
83
+ - null
config.yml CHANGED
@@ -18,7 +18,7 @@
18
  - - learning_starts
19
  - 100000
20
  - - n_timesteps
21
- - 1000000.0
22
  - - optimize_memory_usage
23
  - false
24
  - - policy
 
18
  - - learning_starts
19
  - 100000
20
  - - n_timesteps
21
+ - 100000.0
22
  - - optimize_memory_usage
23
  - false
24
  - - policy
dqn-SpaceInvadersNoFrameskip-v4.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56ea02c51ad747698ee06833c1a0c1e2dbf5ac6dcaeb76391eb84bb3f9eecc36
3
- size 27224715
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5d698f7e0387732d291f05a32b81ac651d1a68f2693feb3d88fc75e596425c5
3
+ size 13719746
dqn-SpaceInvadersNoFrameskip-v4/_stable_baselines3_version CHANGED
@@ -1 +1 @@
1
- 1.6.0
 
1
+ 1.8.0a2
dqn-SpaceInvadersNoFrameskip-v4/data CHANGED
The diff for this file is too large to render. See raw diff
 
dqn-SpaceInvadersNoFrameskip-v4/policy.optimizer.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cb1100db0b6f79a42fc9b7749a142882679b9b6a140c30714db2f8589ac9766
3
- size 13505739
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1e067afe9912f3dd1b7925918b8cbe439229f6008e572c9c7e431ae731419f1
3
+ size 687
dqn-SpaceInvadersNoFrameskip-v4/policy.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bfe7f458228f072d717e250491043cd3061dfa046956c253464621509cda6fb
3
  size 13504937
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9cfcd9e1766805f72358e8889b276d0318eca05e35400bbf4bdc770bb354005
3
  size 13504937
dqn-SpaceInvadersNoFrameskip-v4/system_info.txt CHANGED
@@ -1,7 +1,7 @@
1
- OS: Linux-5.4.0-62-generic-x86_64-with-glibc2.10 #70-Ubuntu SMP Tue Jan 12 12:45:47 UTC 2021
2
- Python: 3.8.8
3
- Stable-Baselines3: 1.6.0
4
- PyTorch: 1.12.0+cu102
5
- GPU Enabled: True
6
- Numpy: 1.20.2
7
- Gym: 0.21.0
 
1
+ - OS: Linux-5.10.147+-x86_64-with-glibc2.29 # 1 SMP Sat Dec 10 16:00:40 UTC 2022
2
+ - Python: 3.8.10
3
+ - Stable-Baselines3: 1.8.0a2
4
+ - PyTorch: 1.13.1+cu116
5
+ - GPU Enabled: True
6
+ - Numpy: 1.21.6
7
+ - Gym: 0.21.0
replay.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fafdd3342b597eebb4e566bcc689b011c9cb6a17f24670ba0bdbf0251585daf3
3
- size 231564
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e38cee50b413c4b8cdda5a1434709b6fd57234bbadab18c33a59a732db44e643
3
+ size 255939
results.json CHANGED
@@ -1 +1 @@
1
- {"mean_reward": 533.5, "std_reward": 199.52506108256176, "is_deterministic": false, "n_eval_episodes": 10, "eval_datetime": "2022-07-22T15:20:39.753290"}
 
1
+ {"mean_reward": 210.0, "std_reward": 145.89379698945393, "is_deterministic": false, "n_eval_episodes": 10, "eval_datetime": "2023-01-29T22:38:26.926819"}
train_eval_metrics.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:667a9a9090b22a281dffd993b2822a2ac16d85076d07e75b8bf3270a1c3b06de
3
- size 37378
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa0570b193c78251d11188d295befaabb70a340db45aaf1effcd1fd2aa78d91b
3
+ size 5845