eshwarprasadS commited on
Commit
0f47df2
·
1 Parent(s): 08fe545

Push PPO CNN Agent for CarRacingv0

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ replay.mp4 filter=lfs diff=lfs merge=lfs -text
PPO_CNN_For_CarRacing.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab25a8b82b56b5c4f3f1083508a76e0c0c2a8446f753fddaceb5ec67e750aa4f
3
+ size 26587488
PPO_CNN_For_CarRacing/_stable_baselines3_version ADDED
@@ -0,0 +1 @@
 
 
1
+ 1.7.0
PPO_CNN_For_CarRacing/data ADDED
The diff for this file is too large to render. See raw diff
 
PPO_CNN_For_CarRacing/policy.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89d1063feaf9930dee6a337e09cf603815cfe33de6d4bacef778cb122a7dd74a
3
+ size 17415600
PPO_CNN_For_CarRacing/policy.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:694f6685bc65c5de10f29439cf5db61fe41bf56ae7053e610e0ae82e7cb6dfce
3
+ size 8709950
PPO_CNN_For_CarRacing/pytorch_variables.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d030ad8db708280fcae77d87e973102039acd23a11bdecc3db8eb6c0ac940ee1
3
+ size 431
PPO_CNN_For_CarRacing/system_info.txt ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - OS: Linux-5.10.147+-x86_64-with-glibc2.29 # 1 SMP Sat Dec 10 16:00:40 UTC 2022
2
+ - Python: 3.8.10
3
+ - Stable-Baselines3: 1.7.0
4
+ - PyTorch: 1.13.1+cu116
5
+ - GPU Enabled: True
6
+ - Numpy: 1.21.6
7
+ - Gym: 0.21.0
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: stable-baselines3
3
+ tags:
4
+ - CarRacing-v0
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - stable-baselines3
8
+ model-index:
9
+ - name: PPO
10
+ results:
11
+ - task:
12
+ type: reinforcement-learning
13
+ name: reinforcement-learning
14
+ dataset:
15
+ name: CarRacing-v0
16
+ type: CarRacing-v0
17
+ metrics:
18
+ - type: mean_reward
19
+ value: 153.28 +/- 117.71
20
+ name: mean_reward
21
+ verified: false
22
+ ---
23
+
24
+ # **PPO** Agent playing **CarRacing-v0**
25
+ This is a trained model of a **PPO** agent playing **CarRacing-v0**
26
+ using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
27
+
28
+ ## Usage (with Stable-baselines3)
29
+ TODO: Add your code
30
+
31
+
32
+ ```python
33
+ from stable_baselines3 import ...
34
+ from huggingface_sb3 import load_from_hub
35
+
36
+ ...
37
+ ```
config.json ADDED
The diff for this file is too large to render. See raw diff
 
replay.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:067b264cd9c1a1373506fd67483def3f0550b1361928b2ef3e561ee28b55f6d6
3
+ size 1084773
results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"mean_reward": 153.27774166315794, "std_reward": 117.7100965950417, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-02-06T11:07:48.947009"}