sugarfreez commited on
Commit
1596fb1
โ€ข
1 Parent(s): df2b142

style(nyz): modify WIP icon

Browse files
Files changed (1) hide show
  1. README.md +13 -8
README.md CHANGED
@@ -26,18 +26,23 @@ If you want to contact us & join us, you can โœ‰๏ธ to our team : <opendilab@p
26
 
27
  # Overview of Model Zoo
28
  <sup>(1): "๐Ÿ”“" means that this algorithm doesn't support this environment.</sup>
29
- <sup>(2): "๐ŸŽฎ" means that the corresponding model is in the upload waitinglist.</sup>
30
  ### Deep Reinforcement Learning
 
 
 
31
  | Algo.\Env. | [LunarLander](https://di-engine-docs.readthedocs.io/en/latest/13_envs/lunarlander.html) | [BipedalWalker](https://di-engine-docs.readthedocs.io/en/latest/13_envs/bipedalwalker.html) | [Pendulum](https://di-engine-docs.readthedocs.io/en/latest/13_envs/pendulum.html) | [Pong](https://di-engine-docs.readthedocs.io/en/latest/13_envs/atari.html) | [SpaceInvaders](https://di-engine-docs.readthedocs.io/en/latest/13_envs/atari.html) | [Qbert](https://di-engine-docs.readthedocs.io/en/latest/13_envs/atari.html) | [Hopper](https://di-engine-docs.readthedocs.io/en/latest/13_envs/mujoco.html) | [Halfcheetah](https://di-engine-docs.readthedocs.io/en/latest/13_envs/mujoco.html) | [Walker2d](https://di-engine-docs.readthedocs.io/en/latest/13_envs/mujoco.html) |
32
  | :-------------: | :-------------: | :------------------------: | :------------: | :--------------: | :------------: | :------------------: | :---------: | :---------: | :---------: |
33
  | [PPO](https://arxiv.org/pdf/1707.06347.pdf) | [โœ…](https://huggingface.co/OpenDILabCommunity/LunarLander-v2-ppo) | | | | | | [โœ…](https://huggingface.co/OpenDILabCommunity/Hopper-v4-PPO) | | |
34
- | [PG](https://proceedings.neurips.cc/paper/1999/file/464d828b85b0bed98e80ade0a5c43b0f-Paper.pdf) | ๐ŸŽฎ | | | | | |๐ŸŽฎ | | |
35
- | [A2C](https://arxiv.org/pdf/1602.01783.pdf) | ๐ŸŽฎ | | | | | | ๐ŸŽฎ | | |
36
- | [IMPALA](https://arxiv.org/pdf/1802.01561.pdf) |๐ŸŽฎ | | | | | | ๐ŸŽฎ | | |
37
- | [DQN](https://storage.googleapis.com/deepmind-media/dqn/DQNNaturePaper.pdf) | ๐ŸŽฎ | | | | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ |
38
- | [DDPG](https://arxiv.org/pdf/1509.02971.pdf) | ๐ŸŽฎ | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ | ๐ŸŽฎ | | |
39
- | [TD3](https://arxiv.org/pdf/1802.09477.pdf) | ๐ŸŽฎ | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ |[โœ…](https://huggingface.co/OpenDILabCommunity/Hopper-v4-TD3) | | |
40
- | [SAC](https://arxiv.org/pdf/1801.01290.pdf) |๐ŸŽฎ | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ | [โœ…](https://huggingface.co/OpenDILabCommunity/Hopper-v4-SAC) | | |
 
 
41
 
42
 
43
  ### Multi-Agent Reinforcement Learning
 
26
 
27
  # Overview of Model Zoo
28
  <sup>(1): "๐Ÿ”“" means that this algorithm doesn't support this environment.</sup>
29
+ <sup>(2): "โณ" means that the corresponding model is in the upload waitinglist (Work In Progress).</sup>
30
  ### Deep Reinforcement Learning
31
+ <details open>
32
+ <summary>(Click to Collapse)</summary>
33
+
34
  | Algo.\Env. | [LunarLander](https://di-engine-docs.readthedocs.io/en/latest/13_envs/lunarlander.html) | [BipedalWalker](https://di-engine-docs.readthedocs.io/en/latest/13_envs/bipedalwalker.html) | [Pendulum](https://di-engine-docs.readthedocs.io/en/latest/13_envs/pendulum.html) | [Pong](https://di-engine-docs.readthedocs.io/en/latest/13_envs/atari.html) | [SpaceInvaders](https://di-engine-docs.readthedocs.io/en/latest/13_envs/atari.html) | [Qbert](https://di-engine-docs.readthedocs.io/en/latest/13_envs/atari.html) | [Hopper](https://di-engine-docs.readthedocs.io/en/latest/13_envs/mujoco.html) | [Halfcheetah](https://di-engine-docs.readthedocs.io/en/latest/13_envs/mujoco.html) | [Walker2d](https://di-engine-docs.readthedocs.io/en/latest/13_envs/mujoco.html) |
35
  | :-------------: | :-------------: | :------------------------: | :------------: | :--------------: | :------------: | :------------------: | :---------: | :---------: | :---------: |
36
  | [PPO](https://arxiv.org/pdf/1707.06347.pdf) | [โœ…](https://huggingface.co/OpenDILabCommunity/LunarLander-v2-ppo) | | | | | | [โœ…](https://huggingface.co/OpenDILabCommunity/Hopper-v4-PPO) | | |
37
+ | [PG](https://proceedings.neurips.cc/paper/1999/file/464d828b85b0bed98e80ade0a5c43b0f-Paper.pdf) | โณ | | | | | | โณ | | |
38
+ | [A2C](https://arxiv.org/pdf/1602.01783.pdf) | โณ | | | | | | โณ | | |
39
+ | [IMPALA](https://arxiv.org/pdf/1802.01561.pdf) | โณ | | | | | | โณ | | |
40
+ | [DQN](https://storage.googleapis.com/deepmind-media/dqn/DQNNaturePaper.pdf) | โณ | | | | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ |
41
+ | [DDPG](https://arxiv.org/pdf/1509.02971.pdf) | โณ | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ | โณ | | |
42
+ | [TD3](https://arxiv.org/pdf/1802.09477.pdf) | โณ | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ |[โœ…](https://huggingface.co/OpenDILabCommunity/Hopper-v4-TD3) | | |
43
+ | [SAC](https://arxiv.org/pdf/1801.01290.pdf) | โณ | | | ๐Ÿ”“ | ๐Ÿ”“ | ๐Ÿ”“ | [โœ…](https://huggingface.co/OpenDILabCommunity/Hopper-v4-SAC) | | |
44
+
45
+ </details>
46
 
47
 
48
  ### Multi-Agent Reinforcement Learning