awacke1 commited on
Commit
ae82371
1 Parent(s): ecad805

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +12 -0
app.py CHANGED
@@ -35,7 +35,11 @@ def load_model(env_name):
35
  model = PPO.load(checkpoint, custom_objects=custom_objects)
36
 
37
  return model
 
 
38
 
 
 
39
 
40
  env_name = st.selectbox(
41
  "Select environment",
@@ -44,6 +48,14 @@ env_name = st.selectbox(
44
  "QbertNoFrameskip-v4",
45
  "SpaceInvadersNoFrameskip-v4",
46
  "PongNoFrameskip-v4",
 
 
 
 
 
 
 
 
47
  ),
48
  )
49
 
 
35
  model = PPO.load(checkpoint, custom_objects=custom_objects)
36
 
37
  return model
38
+
39
+ st.subheader("In game theory and optimization Nash Equilibrium loss minimization starts playing randomly but then by understanding ratios of action success to action-reward with an action (observe, decide/predict, act and then observe outcome the Deep RL agents go from 50% efficiency to 98-99% efficiency based on quality of decision without making mistakes.")
40
 
41
+ st.subheader("list of agent environments https://github.com/DLR-RM/rl-baselines3-zoo/blob/master/benchmark.md")
42
+ st.subheader("https://huggingface.co/sb3")
43
 
44
  env_name = st.selectbox(
45
  "Select environment",
 
48
  "QbertNoFrameskip-v4",
49
  "SpaceInvadersNoFrameskip-v4",
50
  "PongNoFrameskip-v4",
51
+ "AsteroidsNoFrameskip-v4",
52
+ "BeamRiderNoFrameskip-v4",
53
+ "BreakoutNoFrameskip-v4 ",
54
+ "EnduroNoFrameskip-v4",
55
+ "MsPacmanNoFrameskip-v4",
56
+ "RoadRunnerNoFrameskip-v4",
57
+ "Swimmer-v3",
58
+ "Walker2d-v3",
59
  ),
60
  )
61