mus-shd commited on
Commit
7cffbd0
1 Parent(s): ea496e4

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - unity-ml-agents
5
+ - ml-agents
6
+ - deep-reinforcement-learning
7
+ - reinforcement-learning
8
+ - ML-Agents-SoccerTwos
9
+ library_name: ml-agents
10
+ ---
11
+
12
+ # **ppo** Agent playing **SoccerTwos**
13
+ This is a trained model of a **ppo** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
14
+
15
+ ## Usage (with ML-Agents)
16
+ The Documentation: https://github.com/huggingface/ml-agents#get-started
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+
19
+
20
+ ### Resume the training
21
+ ```
22
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
23
+ ```
24
+ ### Watch your Agent play
25
+ You can watch your agent **playing directly in your browser:**.
26
+
27
+ 1. Go to https://huggingface.co/spaces/unity/ML-Agents-SoccerTwos
28
+ 2. Step 1: Write your model_id: chqmatteo/poca-SoccerTwos
29
+ 3. Step 2: Select your *.nn /*.onnx file
30
+ 4. Click on Watch the agent play 👀
31
+