brettgoehre commited on
Commit
85f77bb
1 Parent(s): 6f037fc

First Push

Browse files
README.md ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - SoccerTwos
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-SoccerTwos
8
+ ---
9
+
10
+ # **poca** Agent playing **SoccerTwos**
11
+ This is a trained model of a **poca** agent playing **SoccerTwos**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ ## Usage (with ML-Agents)
15
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
+
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+ - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
+ - A *longer tutorial* to understand how works ML-Agents:
21
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
+
23
+ ### Resume the training
24
+ ```bash
25
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
+ ```
27
+
28
+ ### Watch your Agent play
29
+ You can watch your agent **playing directly in your browser**
30
+
31
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
+ 2. Step 1: Find your model_id: brettgoehre/poca-SoccerTwosMemory
33
+ 3. Step 2: Select your *.nn /*.onnx file
34
+ 4. Click on Watch the agent play 👀
35
+
SoccerTwos.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:382fa58a72199d6dd58f6869e3c1f4b5c0353f94cbfd731a4da94792bbed3b65
3
+ size 2351062
SoccerTwos/SoccerTwos-1499394.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b402ac2672cff38dc844051a6d73e465d3dd949b401c61d19952131be72cbed
3
+ size 2351062
SoccerTwos/SoccerTwos-1499394.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c7629c8bcfd53335d927f2f115d449c19062a9151c10b41fdd0b77b57c32279
3
+ size 36302950
SoccerTwos/SoccerTwos-499360.onnx ADDED
File without changes
SoccerTwos/SoccerTwos-499360.pt ADDED
File without changes
SoccerTwos/SoccerTwos-999356.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ad1afecace049c7dbb1837f6c37b4f33aed734d733e82f7c3e099cca9f2b49f
3
+ size 2351062
SoccerTwos/SoccerTwos-999356.pt ADDED
File without changes
SoccerTwos/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2252c8e3bdc2c4ab2c4dc3266a3ae51e88ff1886224bdc9f6de3441664eb90b5
3
+ size 36301318
SoccerTwos/events.out.tfevents.1725478844.Bretts-MacBook-Pro.local.72766.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50e21b67fe74f262ba283401a667969191f6f2bf26c97580c35179f80946c513
3
+ size 279640
SoccerTwos/events.out.tfevents.1725486116.Bretts-MacBook-Pro.local.3446.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11f398e3557dd9aa0ebc01b8423fcdae2d887d248b0e6499ec45afde51f1b484
3
+ size 424931
SoccerTwos/events.out.tfevents.1725499029.Bretts-MBP.home.48420.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ed43ddcae413951fa86b3bc4f134069b0c69bf306a74a9ad8e8832b54a4c791
3
+ size 874936
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"default_settings": null, "behaviors": {"SoccerTwos": {"trainer_type": "poca", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.9, "num_epoch": 3, "learning_rate_schedule": "constant", "beta_schedule": "linear", "epsilon_schedule": "constant"}, "checkpoint_interval": 500000, "network_settings": {"normalize": true, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": {"sequence_length": 16, "memory_size": 128}, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.995, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "curiosity": {"gamma": 0.9, "strength": 0.05, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0003, "encoding_size": 512}, "rnd": {"gamma": 0.9, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 1, "even_checkpoints": false, "max_steps": 80000000, "time_horizon": 1000, "summary_freq": 10000, "threaded": false, "self_play": {"save_steps": 100000, "team_change": 400000, "swap_steps": 2000, "window": 30, "play_against_latest_model_ratio": 0.9, "initial_elo": 1.0}, "behavioral_cloning": null}}, "env_settings": {"env_path": "./SoccerTwos/SoccerTwos.app", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "sept-run-01", "initialize_from": null, "load_model": false, "resume": true, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ default_settings: null
2
+ behaviors:
3
+ SoccerTwos:
4
+ trainer_type: poca
5
+ hyperparameters:
6
+ batch_size: 2048
7
+ buffer_size: 20480
8
+ learning_rate: 0.0003
9
+ beta: 0.01
10
+ epsilon: 0.2
11
+ lambd: 0.9
12
+ num_epoch: 3
13
+ learning_rate_schedule: constant
14
+ beta_schedule: linear
15
+ epsilon_schedule: constant
16
+ checkpoint_interval: 500000
17
+ network_settings:
18
+ normalize: true
19
+ hidden_units: 512
20
+ num_layers: 2
21
+ vis_encode_type: simple
22
+ memory:
23
+ sequence_length: 16
24
+ memory_size: 128
25
+ goal_conditioning_type: hyper
26
+ deterministic: false
27
+ reward_signals:
28
+ extrinsic:
29
+ gamma: 0.995
30
+ strength: 1.0
31
+ network_settings:
32
+ normalize: false
33
+ hidden_units: 128
34
+ num_layers: 2
35
+ vis_encode_type: simple
36
+ memory: null
37
+ goal_conditioning_type: hyper
38
+ deterministic: false
39
+ curiosity:
40
+ gamma: 0.9
41
+ strength: 0.05
42
+ network_settings:
43
+ normalize: false
44
+ hidden_units: 512
45
+ num_layers: 2
46
+ vis_encode_type: simple
47
+ memory: null
48
+ goal_conditioning_type: hyper
49
+ deterministic: false
50
+ learning_rate: 0.0003
51
+ encoding_size: 512
52
+ rnd:
53
+ gamma: 0.9
54
+ strength: 0.01
55
+ network_settings:
56
+ normalize: false
57
+ hidden_units: 128
58
+ num_layers: 2
59
+ vis_encode_type: simple
60
+ memory: null
61
+ goal_conditioning_type: hyper
62
+ deterministic: false
63
+ learning_rate: 0.0001
64
+ encoding_size: null
65
+ init_path: null
66
+ keep_checkpoints: 1
67
+ even_checkpoints: false
68
+ max_steps: 80000000
69
+ time_horizon: 1000
70
+ summary_freq: 10000
71
+ threaded: false
72
+ self_play:
73
+ save_steps: 100000
74
+ team_change: 400000
75
+ swap_steps: 2000
76
+ window: 30
77
+ play_against_latest_model_ratio: 0.9
78
+ initial_elo: 1.0
79
+ behavioral_cloning: null
80
+ env_settings:
81
+ env_path: ./SoccerTwos/SoccerTwos.app
82
+ env_args: null
83
+ base_port: 5005
84
+ num_envs: 1
85
+ num_areas: 1
86
+ timeout_wait: 60
87
+ seed: -1
88
+ max_lifetime_restarts: 10
89
+ restarts_rate_limit_n: 1
90
+ restarts_rate_limit_period_s: 60
91
+ engine_settings:
92
+ width: 84
93
+ height: 84
94
+ quality_level: 5
95
+ time_scale: 20
96
+ target_frame_rate: -1
97
+ capture_frame_rate: 60
98
+ no_graphics: true
99
+ no_graphics_monitor: false
100
+ environment_parameters: null
101
+ checkpoint_settings:
102
+ run_id: sept-run-01
103
+ initialize_from: null
104
+ load_model: false
105
+ resume: true
106
+ force: false
107
+ train_model: false
108
+ inference: false
109
+ results_dir: results
110
+ torch_settings:
111
+ device: null
112
+ debug: false
run_logs/Player-0.log ADDED
@@ -0,0 +1,342 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Mono path[0] = '/Users/brettgohre/Desktop/huggingface/ml-agents/SoccerTwos/SoccerTwos.app/Contents/Resources/Data/Managed'
2
+ Mono config path = '/Users/brettgohre/Desktop/huggingface/ml-agents/SoccerTwos/SoccerTwos.app/Contents/MonoBleedingEdge/etc'
3
+ New input system (experimental) initialized
4
+ Initialize engine version: 2021.3.17f1 (3e8111cac19d)
5
+ [Subsystems] Discovering subsystems at path /Users/brettgohre/Desktop/huggingface/ml-agents/SoccerTwos/SoccerTwos.app/Contents/Resources/Data/UnitySubsystems
6
+ Forcing GfxDevice: Null
7
+ GfxDevice: creating device client; threaded=0; jobified=0
8
+ NullGfxDevice:
9
+ Version: NULL 1.0 [1.0]
10
+ Renderer: Null Device
11
+ Vendor: Unity Technologies
12
+ Begin MonoManager ReloadAssembly
13
+ - Completed reload, in 0.046 seconds
14
+ ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
15
+ ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
16
+ ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
17
+ WARNING: Shader Unsupported: 'Standard (Specular setup)' - All subshaders removed
18
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
19
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
20
+ ERROR: Shader Standard (Specular setup) shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
21
+ WARNING: Shader Unsupported: 'Standard (Specular setup)' - All subshaders removed
22
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
23
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
24
+ WARNING: Shader Unsupported: 'Autodesk Interactive' - All subshaders removed
25
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
26
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
27
+ ERROR: Shader Autodesk Interactive shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
28
+ WARNING: Shader Unsupported: 'Autodesk Interactive' - All subshaders removed
29
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
30
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
31
+ WARNING: Shader Unsupported: 'Standard' - All subshaders removed
32
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
33
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
34
+ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
35
+ WARNING: Shader Unsupported: 'Standard' - All subshaders removed
36
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
37
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
38
+ UnloadTime: 0.207292 ms
39
+ Registered Communicator in Agent.
40
+ ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
+ TOO LONG
42
+ TOO LONG
43
+ TOO LONG
44
+ TOO LONG
45
+ TOO LONG
46
+ TOO LONG
47
+ TOO LONG
48
+ TOO LONG
49
+ TOO LONG
50
+ TOO LONG
51
+ TOO LONG
52
+ TOO LONG
53
+ TOO LONG
54
+ TOO LONG
55
+ TOO LONG
56
+ TOO LONG
57
+ TOO LONG
58
+ TOO LONG
59
+ TOO LONG
60
+ TOO LONG
61
+ TOO LONG
62
+ TOO LONG
63
+ TOO LONG
64
+ TOO LONG
65
+ TOO LONG
66
+ TOO LONG
67
+ TOO LONG
68
+ TOO LONG
69
+ TOO LONG
70
+ TOO LONG
71
+ TOO LONG
72
+ TOO LONG
73
+ TOO LONG
74
+ TOO LONG
75
+ TOO LONG
76
+ TOO LONG
77
+ TOO LONG
78
+ TOO LONG
79
+ TOO LONG
80
+ TOO LONG
81
+ TOO LONG
82
+ TOO LONG
83
+ TOO LONG
84
+ TOO LONG
85
+ TOO LONG
86
+ TOO LONG
87
+ TOO LONG
88
+ TOO LONG
89
+ TOO LONG
90
+ TOO LONG
91
+ TOO LONG
92
+ TOO LONG
93
+ TOO LONG
94
+ TOO LONG
95
+ TOO LONG
96
+ TOO LONG
97
+ TOO LONG
98
+ TOO LONG
99
+ TOO LONG
100
+ TOO LONG
101
+ TOO LONG
102
+ TOO LONG
103
+ TOO LONG
104
+ TOO LONG
105
+ TOO LONG
106
+ TOO LONG
107
+ TOO LONG
108
+ TOO LONG
109
+ TOO LONG
110
+ TOO LONG
111
+ TOO LONG
112
+ TOO LONG
113
+ TOO LONG
114
+ TOO LONG
115
+ TOO LONG
116
+ TOO LONG
117
+ TOO LONG
118
+ TOO LONG
119
+ TOO LONG
120
+ TOO LONG
121
+ TOO LONG
122
+ TOO LONG
123
+ TOO LONG
124
+ TOO LONG
125
+ TOO LONG
126
+ TOO LONG
127
+ TOO LONG
128
+ TOO LONG
129
+ TOO LONG
130
+ TOO LONG
131
+ TOO LONG
132
+ TOO LONG
133
+ TOO LONG
134
+ TOO LONG
135
+ TOO LONG
136
+ TOO LONG
137
+ TOO LONG
138
+ TOO LONG
139
+ TOO LONG
140
+ TOO LONG
141
+ TOO LONG
142
+ TOO LONG
143
+ TOO LONG
144
+ TOO LONG
145
+ TOO LONG
146
+ TOO LONG
147
+ TOO LONG
148
+ TOO LONG
149
+ TOO LONG
150
+ TOO LONG
151
+ TOO LONG
152
+ TOO LONG
153
+ Default audio device was changed, but the audio system failed to initialize it. Attempting to reset sound system.
154
+ TOO LONG
155
+ TOO LONG
156
+ TOO LONG
157
+ TOO LONG
158
+ TOO LONG
159
+ TOO LONG
160
+ TOO LONG
161
+ TOO LONG
162
+ TOO LONG
163
+ TOO LONG
164
+ TOO LONG
165
+ TOO LONG
166
+ TOO LONG
167
+ TOO LONG
168
+ TOO LONG
169
+ TOO LONG
170
+ TOO LONG
171
+ TOO LONG
172
+ TOO LONG
173
+ TOO LONG
174
+ TOO LONG
175
+ TOO LONG
176
+ TOO LONG
177
+ TOO LONG
178
+ TOO LONG
179
+ TOO LONG
180
+ TOO LONG
181
+ TOO LONG
182
+ TOO LONG
183
+ TOO LONG
184
+ TOO LONG
185
+ TOO LONG
186
+ TOO LONG
187
+ TOO LONG
188
+ TOO LONG
189
+ TOO LONG
190
+ TOO LONG
191
+ TOO LONG
192
+ TOO LONG
193
+ TOO LONG
194
+ TOO LONG
195
+ TOO LONG
196
+ TOO LONG
197
+ TOO LONG
198
+ TOO LONG
199
+ TOO LONG
200
+ TOO LONG
201
+ TOO LONG
202
+ TOO LONG
203
+ TOO LONG
204
+ TOO LONG
205
+ TOO LONG
206
+ TOO LONG
207
+ TOO LONG
208
+ TOO LONG
209
+ TOO LONG
210
+ TOO LONG
211
+ TOO LONG
212
+ TOO LONG
213
+ TOO LONG
214
+ TOO LONG
215
+ TOO LONG
216
+ TOO LONG
217
+ TOO LONG
218
+ TOO LONG
219
+ TOO LONG
220
+ TOO LONG
221
+ TOO LONG
222
+ TOO LONG
223
+ TOO LONG
224
+ TOO LONG
225
+ TOO LONG
226
+ TOO LONG
227
+ TOO LONG
228
+ TOO LONG
229
+ TOO LONG
230
+ TOO LONG
231
+ TOO LONG
232
+ TOO LONG
233
+ TOO LONG
234
+ TOO LONG
235
+ TOO LONG
236
+ TOO LONG
237
+ TOO LONG
238
+ TOO LONG
239
+ TOO LONG
240
+ TOO LONG
241
+ TOO LONG
242
+ TOO LONG
243
+ TOO LONG
244
+ TOO LONG
245
+ TOO LONG
246
+ TOO LONG
247
+ TOO LONG
248
+ TOO LONG
249
+ TOO LONG
250
+ TOO LONG
251
+ TOO LONG
252
+ TOO LONG
253
+ TOO LONG
254
+ TOO LONG
255
+ TOO LONG
256
+ TOO LONG
257
+ TOO LONG
258
+ TOO LONG
259
+ TOO LONG
260
+ TOO LONG
261
+ TOO LONG
262
+ TOO LONG
263
+ TOO LONG
264
+ TOO LONG
265
+ TOO LONG
266
+ TOO LONG
267
+ TOO LONG
268
+ TOO LONG
269
+ TOO LONG
270
+ TOO LONG
271
+ TOO LONG
272
+ TOO LONG
273
+ TOO LONG
274
+ TOO LONG
275
+ TOO LONG
276
+ TOO LONG
277
+ TOO LONG
278
+ TOO LONG
279
+ TOO LONG
280
+ TOO LONG
281
+ TOO LONG
282
+ TOO LONG
283
+ TOO LONG
284
+ TOO LONG
285
+ TOO LONG
286
+ TOO LONG
287
+ TOO LONG
288
+ TOO LONG
289
+ TOO LONG
290
+ TOO LONG
291
+ TOO LONG
292
+ TOO LONG
293
+ TOO LONG
294
+ TOO LONG
295
+ TOO LONG
296
+ TOO LONG
297
+ TOO LONG
298
+ TOO LONG
299
+ TOO LONG
300
+ TOO LONG
301
+ TOO LONG
302
+ TOO LONG
303
+ TOO LONG
304
+ TOO LONG
305
+ TOO LONG
306
+ TOO LONG
307
+ TOO LONG
308
+ TOO LONG
309
+ TOO LONG
310
+ TOO LONG
311
+ TOO LONG
312
+ TOO LONG
313
+ TOO LONG
314
+ TOO LONG
315
+ TOO LONG
316
+ TOO LONG
317
+ TOO LONG
318
+ TOO LONG
319
+ TOO LONG
320
+ TOO LONG
321
+ TOO LONG
322
+ TOO LONG
323
+ TOO LONG
324
+ TOO LONG
325
+ TOO LONG
326
+ TOO LONG
327
+ TOO LONG
328
+ TOO LONG
329
+ TOO LONG
330
+ TOO LONG
331
+ TOO LONG
332
+ TOO LONG
333
+ TOO LONG
334
+ TOO LONG
335
+ TOO LONG
336
+ TOO LONG
337
+ TOO LONG
338
+ TOO LONG
339
+ TOO LONG
340
+ TOO LONG
341
+ TOO LONG
342
+ TOO LONG
run_logs/timers.json ADDED
@@ -0,0 +1,476 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "root",
3
+ "gauges": {
4
+ "SoccerTwos.Policy.Entropy.mean": {
5
+ "value": 2.9194252490997314,
6
+ "min": 2.7066967487335205,
7
+ "max": 2.94661283493042,
8
+ "count": 64
9
+ },
10
+ "SoccerTwos.Policy.Entropy.sum": {
11
+ "value": 57547.7109375,
12
+ "min": 4742.3798828125,
13
+ "max": 93634.375,
14
+ "count": 64
15
+ },
16
+ "SoccerTwos.Environment.EpisodeLength.mean": {
17
+ "value": 604.875,
18
+ "min": 322.6923076923077,
19
+ "max": 999.0,
20
+ "count": 64
21
+ },
22
+ "SoccerTwos.Environment.EpisodeLength.sum": {
23
+ "value": 19356.0,
24
+ "min": 3972.0,
25
+ "max": 31968.0,
26
+ "count": 64
27
+ },
28
+ "SoccerTwos.Self-play.ELO.mean": {
29
+ "value": 4.928913831525834,
30
+ "min": -2.9657419777902905,
31
+ "max": 4.928913831525834,
32
+ "count": 50
33
+ },
34
+ "SoccerTwos.Self-play.ELO.sum": {
35
+ "value": 39.43131065220667,
36
+ "min": -23.725935822322324,
37
+ "max": 98.57827663051667,
38
+ "count": 50
39
+ },
40
+ "SoccerTwos.Step.mean": {
41
+ "value": 1069432.0,
42
+ "min": 439134.0,
43
+ "max": 1069432.0,
44
+ "count": 64
45
+ },
46
+ "SoccerTwos.Step.sum": {
47
+ "value": 1069432.0,
48
+ "min": 439134.0,
49
+ "max": 1069432.0,
50
+ "count": 64
51
+ },
52
+ "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
53
+ "value": -0.07731862366199493,
54
+ "min": -0.17547613382339478,
55
+ "max": -0.012409262359142303,
56
+ "count": 64
57
+ },
58
+ "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
59
+ "value": -1.237097978591919,
60
+ "min": -2.31373929977417,
61
+ "max": -0.21196956932544708,
62
+ "count": 64
63
+ },
64
+ "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
65
+ "value": -0.10046061873435974,
66
+ "min": -0.17158803343772888,
67
+ "max": -0.011743116192519665,
68
+ "count": 64
69
+ },
70
+ "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
71
+ "value": -1.6073698997497559,
72
+ "min": -2.326843738555908,
73
+ "max": -0.18669486045837402,
74
+ "count": 64
75
+ },
76
+ "SoccerTwos.Policy.CuriosityBaselineEstimate.mean": {
77
+ "value": 0.04025515913963318,
78
+ "min": 0.03323616459965706,
79
+ "max": 0.061155594885349274,
80
+ "count": 64
81
+ },
82
+ "SoccerTwos.Policy.CuriosityBaselineEstimate.sum": {
83
+ "value": 0.6440825462341309,
84
+ "min": 0.2457103729248047,
85
+ "max": 1.2533018589019775,
86
+ "count": 64
87
+ },
88
+ "SoccerTwos.Policy.CuriosityValueEstimate.mean": {
89
+ "value": 0.03912646323442459,
90
+ "min": 0.032525066286325455,
91
+ "max": 0.06183749437332153,
92
+ "count": 64
93
+ },
94
+ "SoccerTwos.Policy.CuriosityValueEstimate.sum": {
95
+ "value": 0.6260234117507935,
96
+ "min": 0.24073556065559387,
97
+ "max": 1.2407689094543457,
98
+ "count": 64
99
+ },
100
+ "SoccerTwos.Policy.RndBaselineEstimate.mean": {
101
+ "value": 0.0281881932169199,
102
+ "min": -0.0028001777827739716,
103
+ "max": 0.044277217239141464,
104
+ "count": 64
105
+ },
106
+ "SoccerTwos.Policy.RndBaselineEstimate.sum": {
107
+ "value": 0.4510110914707184,
108
+ "min": -0.03360213339328766,
109
+ "max": 0.6157264113426208,
110
+ "count": 64
111
+ },
112
+ "SoccerTwos.Policy.RndValueEstimate.mean": {
113
+ "value": 0.035632044076919556,
114
+ "min": -0.003357115900143981,
115
+ "max": 0.0452946312725544,
116
+ "count": 64
117
+ },
118
+ "SoccerTwos.Policy.RndValueEstimate.sum": {
119
+ "value": 0.5701127052307129,
120
+ "min": -0.0402853898704052,
121
+ "max": 0.6341248154640198,
122
+ "count": 64
123
+ },
124
+ "SoccerTwos.Environment.CumulativeReward.mean": {
125
+ "value": 0.0,
126
+ "min": 0.0,
127
+ "max": 0.0,
128
+ "count": 64
129
+ },
130
+ "SoccerTwos.Environment.CumulativeReward.sum": {
131
+ "value": 0.0,
132
+ "min": 0.0,
133
+ "max": 0.0,
134
+ "count": 64
135
+ },
136
+ "SoccerTwos.Policy.ExtrinsicReward.mean": {
137
+ "value": -0.29100000113248825,
138
+ "min": -0.5357724140430319,
139
+ "max": 0.2798571416309902,
140
+ "count": 64
141
+ },
142
+ "SoccerTwos.Policy.ExtrinsicReward.sum": {
143
+ "value": -4.656000018119812,
144
+ "min": -15.537400007247925,
145
+ "max": 3.9179999828338623,
146
+ "count": 64
147
+ },
148
+ "SoccerTwos.Policy.CuriosityReward.mean": {
149
+ "value": 2.320769681595266,
150
+ "min": 0.0,
151
+ "max": 5.748547983169556,
152
+ "count": 64
153
+ },
154
+ "SoccerTwos.Policy.CuriosityReward.sum": {
155
+ "value": 37.132314905524254,
156
+ "min": 0.0,
157
+ "max": 61.15827703475952,
158
+ "count": 64
159
+ },
160
+ "SoccerTwos.Policy.RndReward.mean": {
161
+ "value": 0.5761634132359177,
162
+ "min": 0.31020875843562956,
163
+ "max": 4.226163029670715,
164
+ "count": 64
165
+ },
166
+ "SoccerTwos.Policy.RndReward.sum": {
167
+ "value": 9.218614611774683,
168
+ "min": 5.0982994213700294,
169
+ "max": 45.86141597852111,
170
+ "count": 64
171
+ },
172
+ "SoccerTwos.Environment.GroupCumulativeReward.mean": {
173
+ "value": -0.29100000113248825,
174
+ "min": -0.5357724140430319,
175
+ "max": 0.2798571416309902,
176
+ "count": 64
177
+ },
178
+ "SoccerTwos.Environment.GroupCumulativeReward.sum": {
179
+ "value": -4.656000018119812,
180
+ "min": -15.537400007247925,
181
+ "max": 3.9179999828338623,
182
+ "count": 64
183
+ },
184
+ "SoccerTwos.IsTraining.mean": {
185
+ "value": 1.0,
186
+ "min": 1.0,
187
+ "max": 1.0,
188
+ "count": 64
189
+ },
190
+ "SoccerTwos.IsTraining.sum": {
191
+ "value": 1.0,
192
+ "min": 1.0,
193
+ "max": 1.0,
194
+ "count": 64
195
+ },
196
+ "SoccerTwos.Losses.PolicyLoss.mean": {
197
+ "value": 0.05118437986044834,
198
+ "min": 0.034444741702949006,
199
+ "max": 0.07191514666968336,
200
+ "count": 28
201
+ },
202
+ "SoccerTwos.Losses.PolicyLoss.sum": {
203
+ "value": 0.05118437986044834,
204
+ "min": 0.034444741702949006,
205
+ "max": 0.07191514666968336,
206
+ "count": 28
207
+ },
208
+ "SoccerTwos.Losses.ValueLoss.mean": {
209
+ "value": 0.0020198069929998987,
210
+ "min": 2.9689358901426505e-05,
211
+ "max": 0.0020198069929998987,
212
+ "count": 28
213
+ },
214
+ "SoccerTwos.Losses.ValueLoss.sum": {
215
+ "value": 0.0020198069929998987,
216
+ "min": 2.9689358901426505e-05,
217
+ "max": 0.0020198069929998987,
218
+ "count": 28
219
+ },
220
+ "SoccerTwos.Losses.BaselineLoss.mean": {
221
+ "value": 0.0020883430406684056,
222
+ "min": 2.673255736832895e-05,
223
+ "max": 0.0020883430406684056,
224
+ "count": 28
225
+ },
226
+ "SoccerTwos.Losses.BaselineLoss.sum": {
227
+ "value": 0.0020883430406684056,
228
+ "min": 2.673255736832895e-05,
229
+ "max": 0.0020883430406684056,
230
+ "count": 28
231
+ },
232
+ "SoccerTwos.Policy.LearningRate.mean": {
233
+ "value": 0.0003,
234
+ "min": 0.0003,
235
+ "max": 0.0003,
236
+ "count": 28
237
+ },
238
+ "SoccerTwos.Policy.LearningRate.sum": {
239
+ "value": 0.0003,
240
+ "min": 0.0003,
241
+ "max": 0.0003,
242
+ "count": 28
243
+ },
244
+ "SoccerTwos.Policy.Epsilon.mean": {
245
+ "value": 0.20000000000000007,
246
+ "min": 0.19999999999999996,
247
+ "max": 0.20000000000000007,
248
+ "count": 28
249
+ },
250
+ "SoccerTwos.Policy.Epsilon.sum": {
251
+ "value": 0.20000000000000007,
252
+ "min": 0.19999999999999996,
253
+ "max": 0.20000000000000007,
254
+ "count": 28
255
+ },
256
+ "SoccerTwos.Policy.Beta.mean": {
257
+ "value": 0.009868430950750002,
258
+ "min": 0.009868430950750002,
259
+ "max": 0.009941550008500001,
260
+ "count": 28
261
+ },
262
+ "SoccerTwos.Policy.Beta.sum": {
263
+ "value": 0.009868430950750002,
264
+ "min": 0.009868430950750002,
265
+ "max": 0.009941550008500001,
266
+ "count": 28
267
+ },
268
+ "SoccerTwos.Losses.CuriosityForwardLoss.mean": {
269
+ "value": 0.07661932905515036,
270
+ "min": 0.07661932905515036,
271
+ "max": 0.12366530618497304,
272
+ "count": 28
273
+ },
274
+ "SoccerTwos.Losses.CuriosityForwardLoss.sum": {
275
+ "value": 0.07661932905515036,
276
+ "min": 0.07661932905515036,
277
+ "max": 0.12366530618497304,
278
+ "count": 28
279
+ },
280
+ "SoccerTwos.Losses.CuriosityInverseLoss.mean": {
281
+ "value": 2.333798543612162,
282
+ "min": 2.103957215944926,
283
+ "max": 2.3597320238749186,
284
+ "count": 28
285
+ },
286
+ "SoccerTwos.Losses.CuriosityInverseLoss.sum": {
287
+ "value": 2.333798543612162,
288
+ "min": 2.103957215944926,
289
+ "max": 2.3597320238749186,
290
+ "count": 28
291
+ },
292
+ "SoccerTwos.Losses.RNDLoss.mean": {
293
+ "value": 0.09128604084253311,
294
+ "min": 0.09128604084253311,
295
+ "max": 0.4021055996417999,
296
+ "count": 28
297
+ },
298
+ "SoccerTwos.Losses.RNDLoss.sum": {
299
+ "value": 0.09128604084253311,
300
+ "min": 0.09128604084253311,
301
+ "max": 0.4021055996417999,
302
+ "count": 28
303
+ }
304
+ },
305
+ "metadata": {
306
+ "timer_format_version": "0.1.0",
307
+ "start_time_seconds": "1725486114",
308
+ "python_version": "3.10.12 (main, Aug 23 2024, 16:29:11) [Clang 15.0.0 (clang-1500.3.9.4)]",
309
+ "command_line_arguments": "/Users/brettgohre/Desktop/huggingface/ml-agents/games/bin/mlagents-learn ./config/poca/sept/one.yaml --env=./SoccerTwos/SoccerTwos.app --run-id=sept-run-01 --no-graphics --resume",
310
+ "mlagents_version": "1.1.0.dev0",
311
+ "mlagents_envs_version": "1.1.0.dev0",
312
+ "communication_protocol_version": "1.5.0",
313
+ "pytorch_version": "2.4.0",
314
+ "numpy_version": "1.23.5",
315
+ "end_time_seconds": "1725488336"
316
+ },
317
+ "total": 2222.1776464590075,
318
+ "count": 1,
319
+ "self": 0.0701700420177076,
320
+ "children": {
321
+ "run_training.setup": {
322
+ "total": 0.013670291999005713,
323
+ "count": 1,
324
+ "self": 0.013670291999005713
325
+ },
326
+ "TrainerController.start_learning": {
327
+ "total": 2222.093806124991,
328
+ "count": 1,
329
+ "self": 0.3377753770328127,
330
+ "children": {
331
+ "TrainerController._reset_env": {
332
+ "total": 1.7997647919983137,
333
+ "count": 3,
334
+ "self": 1.7997647919983137
335
+ },
336
+ "TrainerController.advance": {
337
+ "total": 2219.8443342889514,
338
+ "count": 40646,
339
+ "self": 0.3316819507308537,
340
+ "children": {
341
+ "env_step": {
342
+ "total": 1791.3922533830773,
343
+ "count": 40646,
344
+ "self": 1712.9993363862886,
345
+ "children": {
346
+ "SubprocessEnvManager._take_step": {
347
+ "total": 78.15483157281415,
348
+ "count": 40646,
349
+ "self": 2.5638017278397456,
350
+ "children": {
351
+ "TorchPolicy.evaluate": {
352
+ "total": 75.5910298449744,
353
+ "count": 80638,
354
+ "self": 75.5910298449744
355
+ }
356
+ }
357
+ },
358
+ "workers": {
359
+ "total": 0.23808542397455312,
360
+ "count": 40645,
361
+ "self": 0.0,
362
+ "children": {
363
+ "worker_root": {
364
+ "total": 2220.0730693208898,
365
+ "count": 40645,
366
+ "is_parallel": true,
367
+ "self": 554.5647922483477,
368
+ "children": {
369
+ "steps_from_proto": {
370
+ "total": 0.004229875979945064,
371
+ "count": 6,
372
+ "is_parallel": true,
373
+ "self": 0.0005952939682174474,
374
+ "children": {
375
+ "_process_rank_one_or_two_observation": {
376
+ "total": 0.003634582011727616,
377
+ "count": 24,
378
+ "is_parallel": true,
379
+ "self": 0.003634582011727616
380
+ }
381
+ }
382
+ },
383
+ "UnityEnvironment.step": {
384
+ "total": 1665.5040471965622,
385
+ "count": 40645,
386
+ "is_parallel": true,
387
+ "self": 4.165579487016657,
388
+ "children": {
389
+ "UnityEnvironment._generate_step_input": {
390
+ "total": 29.249617290261085,
391
+ "count": 40645,
392
+ "is_parallel": true,
393
+ "self": 29.249617290261085
394
+ },
395
+ "communicator.exchange": {
396
+ "total": 1578.8149416175438,
397
+ "count": 40645,
398
+ "is_parallel": true,
399
+ "self": 1578.8149416175438
400
+ },
401
+ "steps_from_proto": {
402
+ "total": 53.27390880174062,
403
+ "count": 81290,
404
+ "is_parallel": true,
405
+ "self": 6.224019605710055,
406
+ "children": {
407
+ "_process_rank_one_or_two_observation": {
408
+ "total": 47.04988919603056,
409
+ "count": 325160,
410
+ "is_parallel": true,
411
+ "self": 47.04988919603056
412
+ }
413
+ }
414
+ }
415
+ }
416
+ }
417
+ }
418
+ }
419
+ }
420
+ }
421
+ }
422
+ },
423
+ "trainer_advance": {
424
+ "total": 428.1203989551432,
425
+ "count": 40645,
426
+ "self": 3.3293685391981853,
427
+ "children": {
428
+ "process_trajectory": {
429
+ "total": 138.76527908295975,
430
+ "count": 40645,
431
+ "self": 138.53822916596255,
432
+ "children": {
433
+ "RLTrainer._checkpoint": {
434
+ "total": 0.22704991699720267,
435
+ "count": 2,
436
+ "self": 0.22704991699720267
437
+ }
438
+ }
439
+ },
440
+ "_update_policy": {
441
+ "total": 286.02575133298524,
442
+ "count": 28,
443
+ "self": 58.20280640778947,
444
+ "children": {
445
+ "TorchPOCAOptimizer.update": {
446
+ "total": 227.82294492519577,
447
+ "count": 858,
448
+ "self": 227.82294492519577
449
+ }
450
+ }
451
+ }
452
+ }
453
+ }
454
+ }
455
+ },
456
+ "trainer_threads": {
457
+ "total": 4.169996827840805e-07,
458
+ "count": 1,
459
+ "self": 4.169996827840805e-07
460
+ },
461
+ "TrainerController._save_models": {
462
+ "total": 0.11193125000863802,
463
+ "count": 1,
464
+ "self": 0.001827208005124703,
465
+ "children": {
466
+ "RLTrainer._checkpoint": {
467
+ "total": 0.11010404200351331,
468
+ "count": 1,
469
+ "self": 0.11010404200351331
470
+ }
471
+ }
472
+ }
473
+ }
474
+ }
475
+ }
476
+ }
run_logs/training_status.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "SoccerTwos": {
3
+ "elo": 4.928913831525834,
4
+ "checkpoints": [
5
+ {
6
+ "steps": 1070432,
7
+ "file_path": "results/sept-run-01/SoccerTwos/SoccerTwos-1070432.onnx",
8
+ "reward": 0.0,
9
+ "creation_time": 1725488336.925156,
10
+ "auxillary_file_paths": [
11
+ "results/sept-run-01/SoccerTwos/SoccerTwos-1070432.pt"
12
+ ]
13
+ }
14
+ ],
15
+ "final_checkpoint": {
16
+ "steps": 1070432,
17
+ "file_path": "results/sept-run-01/SoccerTwos.onnx",
18
+ "reward": 0.0,
19
+ "creation_time": 1725488336.925156,
20
+ "auxillary_file_paths": [
21
+ "results/sept-run-01/SoccerTwos/SoccerTwos-1070432.pt"
22
+ ]
23
+ }
24
+ },
25
+ "metadata": {
26
+ "stats_format_version": "0.3.0",
27
+ "mlagents_version": "1.1.0.dev0",
28
+ "torch_version": "2.4.0"
29
+ }
30
+ }