arshpareek commited on
Commit
907e3f0
1 Parent(s): 014dd76

First Push

Browse files
README.md CHANGED
@@ -1,14 +1,14 @@
1
  ---
2
  library_name: ml-agents
3
  tags:
4
- - SnowballTarget
5
  - deep-reinforcement-learning
6
  - reinforcement-learning
7
- - ML-Agents-SnowballTarget
8
  ---
9
 
10
- # **ppo** Agent playing **SnowballTarget**
11
- This is a trained model of a **ppo** agent playing **SnowballTarget**
12
  using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
 
14
  ## Usage (with ML-Agents)
 
1
  ---
2
  library_name: ml-agents
3
  tags:
4
+ - SoccerTwos
5
  - deep-reinforcement-learning
6
  - reinforcement-learning
7
+ - ML-Agents-SoccerTwos
8
  ---
9
 
10
+ # **poca** Agent playing **SoccerTwos**
11
+ This is a trained model of a **poca** agent playing **SoccerTwos**
12
  using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
 
14
  ## Usage (with ML-Agents)
SoccerTwos.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaa52535e24f0e57357222e3dce1c65b68a572949bc278e9d060c097f4e1694d
3
+ size 1768747
SoccerTwos/SoccerTwos-0.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84818626d1f0277e62f99bf38f065766ab335a9e2c70961b7df1f53f92f7b47c
3
+ size 1766874
SoccerTwos/SoccerTwos-0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c10d075bae7cf465371e0fadedf0dd612e514175b205c917b14bde70a848eced
3
+ size 9479794
SoccerTwos/SoccerTwos-1001986.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaa52535e24f0e57357222e3dce1c65b68a572949bc278e9d060c097f4e1694d
3
+ size 1768747
SoccerTwos/SoccerTwos-1001986.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a13f37a785894378f629314274a008b5dda6d749ac44d87384b4381f6d96931f
3
+ size 28432022
SoccerTwos/SoccerTwos-231634.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca4f17ad36908afe1deee090d5ee489b65e4c4532df16aa912a0869bbc59feac
3
+ size 1768747
SoccerTwos/SoccerTwos-231634.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf09797bb0d54d249de141b7a44529adc4970936318183e71dc010c76b5bb8e1
3
+ size 28431898
SoccerTwos/SoccerTwos-499118.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:593273261f6c5059ec2b1afd377646c0b840d7491d1c0f6bcb6f3f208dbdacac
3
+ size 1768747
SoccerTwos/SoccerTwos-499118.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:161aca3b423c1059b0ea4bf50bb996b6bb05e5eaa4084b9869363ebb057e895c
3
+ size 28431898
SoccerTwos/SoccerTwos-999986.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaa52535e24f0e57357222e3dce1c65b68a572949bc278e9d060c097f4e1694d
3
+ size 1768747
SoccerTwos/SoccerTwos-999986.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd33fe8e5b92c4c162fa0907db24f1ba2a6d45e5ce382d08cd2635b9d9c4eea
3
+ size 28431898
SoccerTwos/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cba22f6ffe35da713afc2dfa4711beafa227c2e55027d724f534efb00bce8eb6
3
+ size 28431030
SoccerTwos/events.out.tfevents.1699443425.5dcb74d24f49.23874.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65d2afc8bcb61575053a3e88199f81a2b267da50899d6466ef19f3c812c1c26a
3
+ size 552734
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SoccerTwos": {"trainer_type": "poca", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "constant", "beta_schedule": "constant", "epsilon_schedule": "constant"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 1000000, "time_horizon": 1000, "summary_freq": 10000, "threaded": false, "self_play": {"save_steps": 50000, "team_change": 200000, "swap_steps": 2000, "window": 10, "play_against_latest_model_ratio": 0.5, "initial_elo": 1200.0}, "behavioral_cloning": null}}, "env_settings": {"env_path": "/content/ml-agents/training-envs-executables/linux/SoccerTwos", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SoccerTwos", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -1,23 +1,22 @@
1
  default_settings: null
2
  behaviors:
3
- SnowballTarget:
4
- trainer_type: ppo
5
  hyperparameters:
6
- batch_size: 128
7
- buffer_size: 2048
8
  learning_rate: 0.0003
9
  beta: 0.005
10
  epsilon: 0.2
11
  lambd: 0.95
12
  num_epoch: 3
13
- shared_critic: false
14
- learning_rate_schedule: linear
15
- beta_schedule: linear
16
- epsilon_schedule: linear
17
- checkpoint_interval: 50000
18
  network_settings:
19
  normalize: false
20
- hidden_units: 256
21
  num_layers: 2
22
  vis_encode_type: simple
23
  memory: null
@@ -36,16 +35,22 @@ behaviors:
36
  goal_conditioning_type: hyper
37
  deterministic: false
38
  init_path: null
39
- keep_checkpoints: 10
40
  even_checkpoints: false
41
- max_steps: 200000
42
- time_horizon: 64
43
  summary_freq: 10000
44
- threaded: true
45
- self_play: null
 
 
 
 
 
 
46
  behavioral_cloning: null
47
  env_settings:
48
- env_path: ./training-envs-executables/linux/SnowballTarget/SnowballTarget
49
  env_args: null
50
  base_port: 5005
51
  num_envs: 1
@@ -65,11 +70,11 @@ engine_settings:
65
  no_graphics: true
66
  environment_parameters: null
67
  checkpoint_settings:
68
- run_id: SnowballTarget1
69
  initialize_from: null
70
  load_model: false
71
  resume: false
72
- force: false
73
  train_model: false
74
  inference: false
75
  results_dir: results
 
1
  default_settings: null
2
  behaviors:
3
+ SoccerTwos:
4
+ trainer_type: poca
5
  hyperparameters:
6
+ batch_size: 2048
7
+ buffer_size: 20480
8
  learning_rate: 0.0003
9
  beta: 0.005
10
  epsilon: 0.2
11
  lambd: 0.95
12
  num_epoch: 3
13
+ learning_rate_schedule: constant
14
+ beta_schedule: constant
15
+ epsilon_schedule: constant
16
+ checkpoint_interval: 500000
 
17
  network_settings:
18
  normalize: false
19
+ hidden_units: 512
20
  num_layers: 2
21
  vis_encode_type: simple
22
  memory: null
 
35
  goal_conditioning_type: hyper
36
  deterministic: false
37
  init_path: null
38
+ keep_checkpoints: 5
39
  even_checkpoints: false
40
+ max_steps: 1000000
41
+ time_horizon: 1000
42
  summary_freq: 10000
43
+ threaded: false
44
+ self_play:
45
+ save_steps: 50000
46
+ team_change: 200000
47
+ swap_steps: 2000
48
+ window: 10
49
+ play_against_latest_model_ratio: 0.5
50
+ initial_elo: 1200.0
51
  behavioral_cloning: null
52
  env_settings:
53
+ env_path: /content/ml-agents/training-envs-executables/linux/SoccerTwos
54
  env_args: null
55
  base_port: 5005
56
  num_envs: 1
 
70
  no_graphics: true
71
  environment_parameters: null
72
  checkpoint_settings:
73
+ run_id: SoccerTwos
74
  initialize_from: null
75
  load_model: false
76
  resume: false
77
+ force: true
78
  train_model: false
79
  inference: false
80
  results_dir: results
run_logs/Player-0.log CHANGED
@@ -1,12 +1,9 @@
1
- Mono path[0] = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/Managed'
2
- Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face
6
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face/SnowballTarget
7
- Unable to load player prefs
8
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
- [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
10
  Forcing GfxDevice: Null
11
  GfxDevice: creating device client; threaded=0; jobified=0
12
  NullGfxDevice:
@@ -34,10 +31,24 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.087 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
42
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
43
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
@@ -45,28 +56,467 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.635880 ms
 
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
  Setting up 1 worker threads for Enlighten.
52
  Memory Statistics:
53
  [ALLOC_TEMP_TLS] TLS Allocator
54
  StackAllocators :
55
  [ALLOC_TEMP_MAIN]
56
- Peak usage frame count: [8.0 KB-16.0 KB]: 5368 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
57
  Initial Block Size 4.0 MB
58
  Current Block Size 4.0 MB
59
- Peak Allocated Bytes 2.0 MB
60
  Overflow Count 0
61
  [ALLOC_TEMP_Loading.AsyncRead]
62
  Initial Block Size 64.0 KB
63
  Current Block Size 64.0 KB
64
- Peak Allocated Bytes 198 B
65
  Overflow Count 0
66
  [ALLOC_TEMP_Loading.PreloadManager]
67
  Initial Block Size 256.0 KB
68
- Current Block Size 304.0 KB
69
- Peak Allocated Bytes 228.9 KB
70
  Overflow Count 4
71
  [ALLOC_TEMP_Background Job.Worker 8]
72
  Initial Block Size 32.0 KB
@@ -81,7 +531,7 @@ Memory Statistics:
81
  [ALLOC_TEMP_Job.Worker 0]
82
  Initial Block Size 256.0 KB
83
  Current Block Size 256.0 KB
84
- Peak Allocated Bytes 3.6 KB
85
  Overflow Count 0
86
  [ALLOC_TEMP_Background Job.Worker 10]
87
  Initial Block Size 32.0 KB
@@ -169,22 +619,22 @@ Memory Statistics:
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
- Peak main deferred allocation count 44
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
- Peak usage frame count: [4.0 MB-8.0 MB]: 5460 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
- Peak Allocated memory 6.5 MB
182
  Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
- Peak usage frame count: [16.0 MB-32.0 MB]: 5460 frames
185
  Requested Block Size 16.0 MB
186
  Peak Block count 1
187
- Peak Allocated memory 17.8 MB
188
  Peak Large allocation bytes 16.0 MB
189
  [ALLOC_TEMP_JOB_1_FRAME]
190
  Initial Block Size 2.0 MB
@@ -213,16 +663,16 @@ Memory Statistics:
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
- Peak usage frame count: [32.0 KB-64.0 KB]: 4888 frames, [64.0 KB-128.0 KB]: 572 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
- Peak Allocated memory 66.7 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
- Peak usage frame count: [32.0 KB-64.0 KB]: 5460 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 1
225
- Peak Allocated memory 39.6 KB
226
  Peak Large allocation bytes 0 B
227
  [ALLOC_CACHEOBJECTS] Dual Thread Allocator
228
  Peak main deferred allocation count 0
@@ -231,16 +681,16 @@ Memory Statistics:
231
  Used Block count 1
232
  Peak Allocated bytes 1.0 MB
233
  [ALLOC_CACHEOBJECTS_MAIN]
234
- Peak usage frame count: [0.5 MB-1.0 MB]: 5460 frames
235
  Requested Block Size 4.0 MB
236
  Peak Block count 1
237
  Peak Allocated memory 0.6 MB
238
  Peak Large allocation bytes 0 B
239
  [ALLOC_CACHEOBJECTS_THREAD]
240
- Peak usage frame count: [0.5 MB-1.0 MB]: 5459 frames, [2.0 MB-4.0 MB]: 1 frames
241
  Requested Block Size 4.0 MB
242
- Peak Block count 1
243
- Peak Allocated memory 2.2 MB
244
  Peak Large allocation bytes 0 B
245
  [ALLOC_TYPETREE] Dual Thread Allocator
246
  Peak main deferred allocation count 0
@@ -249,14 +699,14 @@ Memory Statistics:
249
  Used Block count 1
250
  Peak Allocated bytes 1.0 MB
251
  [ALLOC_TYPETREE_MAIN]
252
- Peak usage frame count: [0-1.0 KB]: 5460 frames
253
  Requested Block Size 2.0 MB
254
  Peak Block count 1
255
  Peak Allocated memory 1.0 KB
256
  Peak Large allocation bytes 0 B
257
  [ALLOC_TYPETREE_THREAD]
258
- Peak usage frame count: [1.0 KB-2.0 KB]: 5460 frames
259
  Requested Block Size 2.0 MB
260
  Peak Block count 1
261
- Peak Allocated memory 1.7 KB
262
  Peak Large allocation bytes 0 B
 
1
+ Mono path[0] = '/content/ml-agents/training-envs-executables/linux/SoccerTwos_Data/Managed'
2
+ Mono config path = '/content/ml-agents/training-envs-executables/linux/SoccerTwos_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
5
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
+ [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SoccerTwos_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
8
  GfxDevice: creating device client; threaded=0; jobified=0
9
  NullGfxDevice:
 
31
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
  FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
+ - Completed reload, in 0.122 seconds
35
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
38
+ WARNING: Shader Unsupported: 'Standard (Specular setup)' - All subshaders removed
39
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
40
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
41
+ ERROR: Shader Standard (Specular setup) shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
42
+ WARNING: Shader Unsupported: 'Standard (Specular setup)' - All subshaders removed
43
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
44
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
45
+ WARNING: Shader Unsupported: 'Autodesk Interactive' - All subshaders removed
46
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
+ ERROR: Shader Autodesk Interactive shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
49
+ WARNING: Shader Unsupported: 'Autodesk Interactive' - All subshaders removed
50
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
51
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
52
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
53
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
54
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
 
56
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
57
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
58
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
59
+ UnloadTime: 0.879396 ms
60
+ Registered Communicator in Agent.
61
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
62
  requesting resize 84 x 84
63
+ TOO LONG
64
+ TOO LONG
65
+ TOO LONG
66
+ TOO LONG
67
+ TOO LONG
68
+ TOO LONG
69
+ TOO LONG
70
+ TOO LONG
71
+ TOO LONG
72
+ TOO LONG
73
+ TOO LONG
74
+ TOO LONG
75
+ TOO LONG
76
+ TOO LONG
77
+ TOO LONG
78
+ TOO LONG
79
+ TOO LONG
80
+ TOO LONG
81
+ TOO LONG
82
+ TOO LONG
83
+ TOO LONG
84
+ TOO LONG
85
+ TOO LONG
86
+ TOO LONG
87
+ TOO LONG
88
+ TOO LONG
89
+ TOO LONG
90
+ TOO LONG
91
+ TOO LONG
92
+ TOO LONG
93
+ TOO LONG
94
+ TOO LONG
95
+ TOO LONG
96
+ TOO LONG
97
+ TOO LONG
98
+ TOO LONG
99
+ TOO LONG
100
+ TOO LONG
101
+ TOO LONG
102
+ TOO LONG
103
+ TOO LONG
104
+ TOO LONG
105
+ TOO LONG
106
+ TOO LONG
107
+ TOO LONG
108
+ TOO LONG
109
+ TOO LONG
110
+ TOO LONG
111
+ TOO LONG
112
+ TOO LONG
113
+ TOO LONG
114
+ TOO LONG
115
+ TOO LONG
116
+ TOO LONG
117
+ TOO LONG
118
+ TOO LONG
119
+ TOO LONG
120
+ TOO LONG
121
+ TOO LONG
122
+ TOO LONG
123
+ TOO LONG
124
+ TOO LONG
125
+ TOO LONG
126
+ TOO LONG
127
+ TOO LONG
128
+ TOO LONG
129
+ TOO LONG
130
+ TOO LONG
131
+ TOO LONG
132
+ TOO LONG
133
+ TOO LONG
134
+ TOO LONG
135
+ TOO LONG
136
+ TOO LONG
137
+ TOO LONG
138
+ TOO LONG
139
+ TOO LONG
140
+ TOO LONG
141
+ TOO LONG
142
+ TOO LONG
143
+ TOO LONG
144
+ TOO LONG
145
+ TOO LONG
146
+ TOO LONG
147
+ TOO LONG
148
+ TOO LONG
149
+ TOO LONG
150
+ TOO LONG
151
+ TOO LONG
152
+ TOO LONG
153
+ TOO LONG
154
+ TOO LONG
155
+ TOO LONG
156
+ TOO LONG
157
+ TOO LONG
158
+ TOO LONG
159
+ TOO LONG
160
+ TOO LONG
161
+ TOO LONG
162
+ TOO LONG
163
+ TOO LONG
164
+ TOO LONG
165
+ TOO LONG
166
+ TOO LONG
167
+ TOO LONG
168
+ TOO LONG
169
+ TOO LONG
170
+ TOO LONG
171
+ TOO LONG
172
+ TOO LONG
173
+ TOO LONG
174
+ TOO LONG
175
+ TOO LONG
176
+ TOO LONG
177
+ TOO LONG
178
+ TOO LONG
179
+ TOO LONG
180
+ TOO LONG
181
+ TOO LONG
182
+ TOO LONG
183
+ TOO LONG
184
+ TOO LONG
185
+ TOO LONG
186
+ TOO LONG
187
+ TOO LONG
188
+ TOO LONG
189
+ TOO LONG
190
+ TOO LONG
191
+ TOO LONG
192
+ TOO LONG
193
+ TOO LONG
194
+ TOO LONG
195
+ TOO LONG
196
+ TOO LONG
197
+ TOO LONG
198
+ TOO LONG
199
+ TOO LONG
200
+ TOO LONG
201
+ TOO LONG
202
+ TOO LONG
203
+ TOO LONG
204
+ TOO LONG
205
+ TOO LONG
206
+ TOO LONG
207
+ TOO LONG
208
+ TOO LONG
209
+ TOO LONG
210
+ TOO LONG
211
+ TOO LONG
212
+ TOO LONG
213
+ TOO LONG
214
+ TOO LONG
215
+ TOO LONG
216
+ TOO LONG
217
+ TOO LONG
218
+ TOO LONG
219
+ TOO LONG
220
+ TOO LONG
221
+ TOO LONG
222
+ TOO LONG
223
+ TOO LONG
224
+ TOO LONG
225
+ TOO LONG
226
+ TOO LONG
227
+ TOO LONG
228
+ TOO LONG
229
+ TOO LONG
230
+ TOO LONG
231
+ TOO LONG
232
+ TOO LONG
233
+ TOO LONG
234
+ TOO LONG
235
+ TOO LONG
236
+ TOO LONG
237
+ TOO LONG
238
+ TOO LONG
239
+ TOO LONG
240
+ TOO LONG
241
+ TOO LONG
242
+ TOO LONG
243
+ TOO LONG
244
+ TOO LONG
245
+ TOO LONG
246
+ TOO LONG
247
+ TOO LONG
248
+ TOO LONG
249
+ TOO LONG
250
+ TOO LONG
251
+ TOO LONG
252
+ TOO LONG
253
+ TOO LONG
254
+ TOO LONG
255
+ TOO LONG
256
+ TOO LONG
257
+ TOO LONG
258
+ TOO LONG
259
+ TOO LONG
260
+ TOO LONG
261
+ TOO LONG
262
+ TOO LONG
263
+ TOO LONG
264
+ TOO LONG
265
+ TOO LONG
266
+ TOO LONG
267
+ TOO LONG
268
+ TOO LONG
269
+ TOO LONG
270
+ TOO LONG
271
+ TOO LONG
272
+ TOO LONG
273
+ TOO LONG
274
+ TOO LONG
275
+ TOO LONG
276
+ TOO LONG
277
+ TOO LONG
278
+ TOO LONG
279
+ TOO LONG
280
+ TOO LONG
281
+ TOO LONG
282
+ TOO LONG
283
+ TOO LONG
284
+ TOO LONG
285
+ TOO LONG
286
+ TOO LONG
287
+ TOO LONG
288
+ TOO LONG
289
+ TOO LONG
290
+ TOO LONG
291
+ TOO LONG
292
+ TOO LONG
293
+ TOO LONG
294
+ TOO LONG
295
+ TOO LONG
296
+ TOO LONG
297
+ TOO LONG
298
+ TOO LONG
299
+ TOO LONG
300
+ TOO LONG
301
+ TOO LONG
302
+ TOO LONG
303
+ TOO LONG
304
+ TOO LONG
305
+ TOO LONG
306
+ TOO LONG
307
+ TOO LONG
308
+ TOO LONG
309
+ TOO LONG
310
+ TOO LONG
311
+ TOO LONG
312
+ TOO LONG
313
+ TOO LONG
314
+ TOO LONG
315
+ TOO LONG
316
+ TOO LONG
317
+ TOO LONG
318
+ TOO LONG
319
+ TOO LONG
320
+ TOO LONG
321
+ TOO LONG
322
+ TOO LONG
323
+ TOO LONG
324
+ TOO LONG
325
+ TOO LONG
326
+ TOO LONG
327
+ TOO LONG
328
+ TOO LONG
329
+ TOO LONG
330
+ TOO LONG
331
+ TOO LONG
332
+ TOO LONG
333
+ TOO LONG
334
+ TOO LONG
335
+ TOO LONG
336
+ TOO LONG
337
+ TOO LONG
338
+ TOO LONG
339
+ TOO LONG
340
+ TOO LONG
341
+ TOO LONG
342
+ TOO LONG
343
+ TOO LONG
344
+ TOO LONG
345
+ TOO LONG
346
+ TOO LONG
347
+ TOO LONG
348
+ TOO LONG
349
+ TOO LONG
350
+ TOO LONG
351
+ TOO LONG
352
+ TOO LONG
353
+ TOO LONG
354
+ TOO LONG
355
+ TOO LONG
356
+ TOO LONG
357
+ TOO LONG
358
+ TOO LONG
359
+ TOO LONG
360
+ TOO LONG
361
+ TOO LONG
362
+ TOO LONG
363
+ TOO LONG
364
+ TOO LONG
365
+ TOO LONG
366
+ TOO LONG
367
+ TOO LONG
368
+ TOO LONG
369
+ TOO LONG
370
+ TOO LONG
371
+ TOO LONG
372
+ TOO LONG
373
+ TOO LONG
374
+ TOO LONG
375
+ TOO LONG
376
+ TOO LONG
377
+ TOO LONG
378
+ TOO LONG
379
+ TOO LONG
380
+ TOO LONG
381
+ TOO LONG
382
+ TOO LONG
383
+ TOO LONG
384
+ TOO LONG
385
+ TOO LONG
386
+ TOO LONG
387
+ TOO LONG
388
+ TOO LONG
389
+ TOO LONG
390
+ TOO LONG
391
+ TOO LONG
392
+ TOO LONG
393
+ TOO LONG
394
+ TOO LONG
395
+ TOO LONG
396
+ TOO LONG
397
+ TOO LONG
398
+ TOO LONG
399
+ TOO LONG
400
+ TOO LONG
401
+ TOO LONG
402
+ TOO LONG
403
+ TOO LONG
404
+ TOO LONG
405
+ TOO LONG
406
+ TOO LONG
407
+ TOO LONG
408
+ TOO LONG
409
+ TOO LONG
410
+ TOO LONG
411
+ TOO LONG
412
+ TOO LONG
413
+ TOO LONG
414
+ TOO LONG
415
+ TOO LONG
416
+ TOO LONG
417
+ TOO LONG
418
+ TOO LONG
419
+ TOO LONG
420
+ TOO LONG
421
+ TOO LONG
422
+ TOO LONG
423
+ TOO LONG
424
+ TOO LONG
425
+ TOO LONG
426
+ TOO LONG
427
+ TOO LONG
428
+ TOO LONG
429
+ TOO LONG
430
+ TOO LONG
431
+ TOO LONG
432
+ TOO LONG
433
+ TOO LONG
434
+ TOO LONG
435
+ TOO LONG
436
+ TOO LONG
437
+ TOO LONG
438
+ TOO LONG
439
+ TOO LONG
440
+ TOO LONG
441
+ TOO LONG
442
+ TOO LONG
443
+ TOO LONG
444
+ TOO LONG
445
+ TOO LONG
446
+ TOO LONG
447
+ TOO LONG
448
+ TOO LONG
449
+ TOO LONG
450
+ TOO LONG
451
+ TOO LONG
452
+ TOO LONG
453
+ TOO LONG
454
+ TOO LONG
455
+ TOO LONG
456
+ TOO LONG
457
+ TOO LONG
458
+ TOO LONG
459
+ TOO LONG
460
+ TOO LONG
461
+ TOO LONG
462
+ TOO LONG
463
+ TOO LONG
464
+ TOO LONG
465
+ TOO LONG
466
+ TOO LONG
467
+ TOO LONG
468
+ TOO LONG
469
+ TOO LONG
470
+ TOO LONG
471
+ TOO LONG
472
+ TOO LONG
473
+ TOO LONG
474
+ TOO LONG
475
+ TOO LONG
476
+ TOO LONG
477
+ TOO LONG
478
+ TOO LONG
479
+ TOO LONG
480
+ TOO LONG
481
+ TOO LONG
482
+ TOO LONG
483
+ TOO LONG
484
+ TOO LONG
485
+ TOO LONG
486
+ TOO LONG
487
+ TOO LONG
488
+ TOO LONG
489
+ TOO LONG
490
+ TOO LONG
491
+ TOO LONG
492
+ TOO LONG
493
+ TOO LONG
494
+ TOO LONG
495
+ TOO LONG
496
+ TOO LONG
497
+ TOO LONG
498
+ TOO LONG
499
+ TOO LONG
500
+ TOO LONG
501
  Setting up 1 worker threads for Enlighten.
502
  Memory Statistics:
503
  [ALLOC_TEMP_TLS] TLS Allocator
504
  StackAllocators :
505
  [ALLOC_TEMP_MAIN]
506
+ Peak usage frame count: [8.0 KB-16.0 KB]: 19356 frames, [2.0 MB-4.0 MB]: 1 frames
507
  Initial Block Size 4.0 MB
508
  Current Block Size 4.0 MB
509
+ Peak Allocated Bytes 2.1 MB
510
  Overflow Count 0
511
  [ALLOC_TEMP_Loading.AsyncRead]
512
  Initial Block Size 64.0 KB
513
  Current Block Size 64.0 KB
514
+ Peak Allocated Bytes 128 B
515
  Overflow Count 0
516
  [ALLOC_TEMP_Loading.PreloadManager]
517
  Initial Block Size 256.0 KB
518
+ Current Block Size 300.0 KB
519
+ Peak Allocated Bytes 249.6 KB
520
  Overflow Count 4
521
  [ALLOC_TEMP_Background Job.Worker 8]
522
  Initial Block Size 32.0 KB
 
531
  [ALLOC_TEMP_Job.Worker 0]
532
  Initial Block Size 256.0 KB
533
  Current Block Size 256.0 KB
534
+ Peak Allocated Bytes 0.7 KB
535
  Overflow Count 0
536
  [ALLOC_TEMP_Background Job.Worker 10]
537
  Initial Block Size 32.0 KB
 
619
  Peak Allocated Bytes 0 B
620
  Overflow Count 0
621
  [ALLOC_DEFAULT] Dual Thread Allocator
622
+ Peak main deferred allocation count 4
623
  [ALLOC_BUCKET]
624
  Large Block size 4.0 MB
625
  Used Block count 1
626
  Peak Allocated bytes 1.0 MB
627
  [ALLOC_DEFAULT_MAIN]
628
+ Peak usage frame count: [4.0 MB-8.0 MB]: 19357 frames
629
  Requested Block Size 16.0 MB
630
  Peak Block count 1
631
+ Peak Allocated memory 4.6 MB
632
  Peak Large allocation bytes 0 B
633
  [ALLOC_DEFAULT_THREAD]
634
+ Peak usage frame count: [16.0 MB-32.0 MB]: 19357 frames
635
  Requested Block Size 16.0 MB
636
  Peak Block count 1
637
+ Peak Allocated memory 17.6 MB
638
  Peak Large allocation bytes 16.0 MB
639
  [ALLOC_TEMP_JOB_1_FRAME]
640
  Initial Block Size 2.0 MB
 
663
  Used Block count 1
664
  Peak Allocated bytes 1.0 MB
665
  [ALLOC_GFX_MAIN]
666
+ Peak usage frame count: [32.0 KB-64.0 KB]: 19356 frames, [64.0 KB-128.0 KB]: 1 frames
667
  Requested Block Size 16.0 MB
668
  Peak Block count 1
669
+ Peak Allocated memory 65.6 KB
670
  Peak Large allocation bytes 0 B
671
  [ALLOC_GFX_THREAD]
672
+ Peak usage frame count: [128.0 KB-256.0 KB]: 19357 frames
673
  Requested Block Size 16.0 MB
674
  Peak Block count 1
675
+ Peak Allocated memory 173.5 KB
676
  Peak Large allocation bytes 0 B
677
  [ALLOC_CACHEOBJECTS] Dual Thread Allocator
678
  Peak main deferred allocation count 0
 
681
  Used Block count 1
682
  Peak Allocated bytes 1.0 MB
683
  [ALLOC_CACHEOBJECTS_MAIN]
684
+ Peak usage frame count: [0.5 MB-1.0 MB]: 19357 frames
685
  Requested Block Size 4.0 MB
686
  Peak Block count 1
687
  Peak Allocated memory 0.6 MB
688
  Peak Large allocation bytes 0 B
689
  [ALLOC_CACHEOBJECTS_THREAD]
690
+ Peak usage frame count: [0.5 MB-1.0 MB]: 19356 frames, [4.0 MB-8.0 MB]: 1 frames
691
  Requested Block Size 4.0 MB
692
+ Peak Block count 2
693
+ Peak Allocated memory 4.5 MB
694
  Peak Large allocation bytes 0 B
695
  [ALLOC_TYPETREE] Dual Thread Allocator
696
  Peak main deferred allocation count 0
 
699
  Used Block count 1
700
  Peak Allocated bytes 1.0 MB
701
  [ALLOC_TYPETREE_MAIN]
702
+ Peak usage frame count: [0-1.0 KB]: 19357 frames
703
  Requested Block Size 2.0 MB
704
  Peak Block count 1
705
  Peak Allocated memory 1.0 KB
706
  Peak Large allocation bytes 0 B
707
  [ALLOC_TYPETREE_THREAD]
708
+ Peak usage frame count: [1.0 KB-2.0 KB]: 19356 frames, [2.0 KB-4.0 KB]: 1 frames
709
  Requested Block Size 2.0 MB
710
  Peak Block count 1
711
+ Peak Allocated memory 2.2 KB
712
  Peak Large allocation bytes 0 B
run_logs/timers.json CHANGED
@@ -1,214 +1,262 @@
1
  {
2
  "name": "root",
3
  "gauges": {
4
- "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.8524067401885986,
6
- "min": 0.8524067401885986,
7
- "max": 2.8588712215423584,
8
- "count": 20
9
- },
10
- "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 8138.779296875,
12
- "min": 8138.779296875,
13
- "max": 29309.1484375,
14
- "count": 20
15
- },
16
- "SnowballTarget.Step.mean": {
17
- "value": 199984.0,
18
- "min": 9952.0,
19
- "max": 199984.0,
20
- "count": 20
21
- },
22
- "SnowballTarget.Step.sum": {
23
- "value": 199984.0,
24
- "min": 9952.0,
25
- "max": 199984.0,
26
- "count": 20
27
- },
28
- "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 13.148093223571777,
30
- "min": 0.39425939321517944,
31
- "max": 13.148093223571777,
32
- "count": 20
33
- },
34
- "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2563.878173828125,
36
- "min": 76.48632049560547,
37
- "max": 2652.8505859375,
38
- "count": 20
39
- },
40
- "SnowballTarget.Environment.EpisodeLength.mean": {
41
- "value": 199.0,
42
- "min": 199.0,
43
- "max": 199.0,
44
- "count": 20
45
- },
46
- "SnowballTarget.Environment.EpisodeLength.sum": {
47
- "value": 8756.0,
48
- "min": 8756.0,
49
- "max": 10945.0,
50
- "count": 20
51
- },
52
- "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.07091618571640164,
54
- "min": 0.05967896982689911,
55
- "max": 0.07549933638702165,
56
- "count": 20
57
- },
58
- "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.28366474286560656,
60
- "min": 0.2556595531584439,
61
- "max": 0.372281089425087,
62
- "count": 20
63
- },
64
- "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.19979606955951335,
66
- "min": 0.11655839364610467,
67
- "max": 0.2877348328922309,
68
- "count": 20
69
- },
70
- "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.7991842782380534,
72
- "min": 0.46623357458441866,
73
- "max": 1.4386741644611545,
74
- "count": 20
75
- },
76
- "SnowballTarget.Policy.LearningRate.mean": {
77
- "value": 8.082097306000005e-06,
78
- "min": 8.082097306000005e-06,
79
- "max": 0.000291882002706,
80
- "count": 20
81
- },
82
- "SnowballTarget.Policy.LearningRate.sum": {
83
- "value": 3.232838922400002e-05,
84
- "min": 3.232838922400002e-05,
85
- "max": 0.00138516003828,
86
- "count": 20
87
- },
88
- "SnowballTarget.Policy.Epsilon.mean": {
89
- "value": 0.10269400000000001,
90
- "min": 0.10269400000000001,
91
- "max": 0.19729400000000002,
92
- "count": 20
93
- },
94
- "SnowballTarget.Policy.Epsilon.sum": {
95
- "value": 0.41077600000000003,
96
- "min": 0.41077600000000003,
97
- "max": 0.96172,
98
- "count": 20
99
- },
100
- "SnowballTarget.Policy.Beta.mean": {
101
- "value": 0.0001444306000000001,
102
- "min": 0.0001444306000000001,
103
- "max": 0.0048649706,
104
- "count": 20
105
- },
106
- "SnowballTarget.Policy.Beta.sum": {
107
- "value": 0.0005777224000000004,
108
- "min": 0.0005777224000000004,
109
- "max": 0.023089828,
110
- "count": 20
111
- },
112
- "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 26.386363636363637,
114
- "min": 3.159090909090909,
115
- "max": 26.386363636363637,
116
- "count": 20
117
- },
118
- "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1161.0,
120
- "min": 139.0,
121
- "max": 1420.0,
122
- "count": 20
123
- },
124
- "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 26.386363636363637,
126
- "min": 3.159090909090909,
127
- "max": 26.386363636363637,
128
- "count": 20
129
- },
130
- "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1161.0,
132
- "min": 139.0,
133
- "max": 1420.0,
134
- "count": 20
135
- },
136
- "SnowballTarget.IsTraining.mean": {
137
  "value": 1.0,
138
  "min": 1.0,
139
  "max": 1.0,
140
- "count": 20
141
  },
142
- "SnowballTarget.IsTraining.sum": {
143
  "value": 1.0,
144
  "min": 1.0,
145
  "max": 1.0,
146
- "count": 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
147
  }
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1698484795",
152
  "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
153
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.1.0.dev0",
155
  "mlagents_envs_version": "1.1.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.1.0+cu118",
158
  "numpy_version": "1.23.5",
159
- "end_time_seconds": "1698485309"
160
  },
161
- "total": 513.7411729779999,
162
  "count": 1,
163
- "self": 0.42868975699980183,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.04582125199999609,
167
  "count": 1,
168
- "self": 0.04582125199999609
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 513.2666619690001,
172
  "count": 1,
173
- "self": 0.6647855600027697,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 10.031701074000011,
177
- "count": 1,
178
- "self": 10.031701074000011
179
  },
180
  "TrainerController.advance": {
181
- "total": 502.4853038939972,
182
- "count": 18200,
183
- "self": 0.3250254600028484,
184
  "children": {
185
  "env_step": {
186
- "total": 502.16027843399434,
187
- "count": 18200,
188
- "self": 345.4591988499914,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 156.36378232900233,
192
- "count": 18200,
193
- "self": 1.632031328009532,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 154.7317510009928,
197
- "count": 18200,
198
- "self": 154.7317510009928
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.3372972550005784,
204
- "count": 18200,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 511.7454815879949,
209
- "count": 18200,
210
  "is_parallel": true,
211
- "self": 248.06104059100517,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +265,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.006495056000005661,
221
- "count": 1,
222
  "is_parallel": true,
223
- "self": 0.00447220099988499,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0020228550001206713,
227
- "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0020228550001206713
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.07582938600000944,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0006761670000514641,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.0004288309999651574,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.0004288309999651574
244
  },
245
  "communicator.exchange": {
246
- "total": 0.07285548900000549,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.07285548900000549
250
  },
251
  "steps_from_proto": {
252
- "total": 0.001868898999987323,
253
- "count": 1,
254
  "is_parallel": true,
255
- "self": 0.00034124900003007497,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.001527649999957248,
259
- "count": 10,
260
  "is_parallel": true,
261
- "self": 0.001527649999957248
262
  }
263
  }
264
  }
@@ -267,91 +315,85 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 263.68444099698974,
271
- "count": 18199,
272
  "is_parallel": true,
273
- "self": 11.489354049983206,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 5.773022557992931,
277
- "count": 18199,
278
  "is_parallel": true,
279
- "self": 5.773022557992931
280
  },
281
  "communicator.exchange": {
282
- "total": 209.5062561730074,
283
- "count": 18199,
284
  "is_parallel": true,
285
- "self": 209.5062561730074
286
  },
287
  "steps_from_proto": {
288
- "total": 36.91580821600621,
289
- "count": 18199,
290
  "is_parallel": true,
291
- "self": 7.3905490550027935,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 29.52525916100342,
295
- "count": 181990,
296
  "is_parallel": true,
297
- "self": 29.52525916100342
298
  }
299
  }
300
  }
301
  }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
302
  }
303
  }
304
  }
305
  }
306
  }
307
  }
308
- }
309
- }
310
- },
311
- "trainer_threads": {
312
- "total": 0.00010489100009181129,
313
- "count": 1,
314
- "self": 0.00010489100009181129,
315
- "children": {
316
- "thread_root": {
317
- "total": 0.0,
318
- "count": 0,
319
- "is_parallel": true,
320
- "self": 0.0,
321
  "children": {
322
- "trainer_advance": {
323
- "total": 498.2910548509996,
324
- "count": 532691,
325
- "is_parallel": true,
326
- "self": 11.420711338001809,
327
  "children": {
328
- "process_trajectory": {
329
- "total": 282.6814876759978,
330
- "count": 532691,
331
- "is_parallel": true,
332
- "self": 281.8395120169978,
333
- "children": {
334
- "RLTrainer._checkpoint": {
335
- "total": 0.8419756590000134,
336
- "count": 4,
337
- "is_parallel": true,
338
- "self": 0.8419756590000134
339
- }
340
- }
341
- },
342
- "_update_policy": {
343
- "total": 204.188855837,
344
- "count": 90,
345
- "is_parallel": true,
346
- "self": 62.57992527299717,
347
- "children": {
348
- "TorchPPOOptimizer.update": {
349
- "total": 141.60893056400283,
350
- "count": 4584,
351
- "is_parallel": true,
352
- "self": 141.60893056400283
353
- }
354
- }
355
  }
356
  }
357
  }
@@ -359,15 +401,20 @@
359
  }
360
  }
361
  },
 
 
 
 
 
362
  "TrainerController._save_models": {
363
- "total": 0.0847665500000403,
364
  "count": 1,
365
- "self": 0.0008246569999528219,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
- "total": 0.08394189300008748,
369
  "count": 1,
370
- "self": 0.08394189300008748
371
  }
372
  }
373
  }
 
1
  {
2
  "name": "root",
3
  "gauges": {
4
+ "SoccerTwos.Policy.Entropy.mean": {
5
+ "value": 3.149009943008423,
6
+ "min": 3.0920655727386475,
7
+ "max": 3.295755386352539,
8
+ "count": 100
9
+ },
10
+ "SoccerTwos.Policy.Entropy.sum": {
11
+ "value": 57639.4765625,
12
+ "min": 17742.544921875,
13
+ "max": 123030.3515625,
14
+ "count": 100
15
+ },
16
+ "SoccerTwos.Environment.EpisodeLength.mean": {
17
+ "value": 999.0,
18
+ "min": 419.9230769230769,
19
+ "max": 999.0,
20
+ "count": 100
21
+ },
22
+ "SoccerTwos.Environment.EpisodeLength.sum": {
23
+ "value": 19980.0,
24
+ "min": 11988.0,
25
+ "max": 28196.0,
26
+ "count": 100
27
+ },
28
+ "SoccerTwos.Self-play.ELO.mean": {
29
+ "value": 1197.1405668605373,
30
+ "min": 1194.6627015140439,
31
+ "max": 1204.034278640044,
32
+ "count": 73
33
+ },
34
+ "SoccerTwos.Self-play.ELO.sum": {
35
+ "value": 2394.2811337210746,
36
+ "min": 2392.647101577426,
37
+ "max": 16821.98242286591,
38
+ "count": 73
39
+ },
40
+ "SoccerTwos.Step.mean": {
41
+ "value": 999986.0,
42
+ "min": 9116.0,
43
+ "max": 999986.0,
44
+ "count": 100
45
+ },
46
+ "SoccerTwos.Step.sum": {
47
+ "value": 999986.0,
48
+ "min": 9116.0,
49
+ "max": 999986.0,
50
+ "count": 100
51
+ },
52
+ "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
53
+ "value": 0.002587884897366166,
54
+ "min": -0.03601016849279404,
55
+ "max": 0.005797781050205231,
56
+ "count": 100
57
+ },
58
+ "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
59
+ "value": 0.025878848508000374,
60
+ "min": -0.5041423439979553,
61
+ "max": 0.07835473120212555,
62
+ "count": 100
63
+ },
64
+ "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
65
+ "value": 0.0025931489653885365,
66
+ "min": -0.039575692266225815,
67
+ "max": 0.005782643333077431,
68
+ "count": 100
69
+ },
70
+ "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
71
+ "value": 0.02593148872256279,
72
+ "min": -0.5540596842765808,
73
+ "max": 0.07517436146736145,
74
+ "count": 100
75
+ },
76
+ "SoccerTwos.Environment.CumulativeReward.mean": {
77
+ "value": 0.0,
78
+ "min": 0.0,
79
+ "max": 0.0,
80
+ "count": 100
81
+ },
82
+ "SoccerTwos.Environment.CumulativeReward.sum": {
83
+ "value": 0.0,
84
+ "min": 0.0,
85
+ "max": 0.0,
86
+ "count": 100
87
+ },
88
+ "SoccerTwos.Policy.ExtrinsicReward.mean": {
89
+ "value": 0.0,
90
+ "min": -0.5625,
91
+ "max": 0.3017142776932035,
92
+ "count": 100
93
+ },
94
+ "SoccerTwos.Policy.ExtrinsicReward.sum": {
95
+ "value": 0.0,
96
+ "min": -9.0,
97
+ "max": 4.223999887704849,
98
+ "count": 100
99
+ },
100
+ "SoccerTwos.Environment.GroupCumulativeReward.mean": {
101
+ "value": 0.0,
102
+ "min": -0.5625,
103
+ "max": 0.3017142776932035,
104
+ "count": 100
105
+ },
106
+ "SoccerTwos.Environment.GroupCumulativeReward.sum": {
107
+ "value": 0.0,
108
+ "min": -9.0,
109
+ "max": 4.223999887704849,
110
+ "count": 100
111
+ },
112
+ "SoccerTwos.IsTraining.mean": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  "value": 1.0,
114
  "min": 1.0,
115
  "max": 1.0,
116
+ "count": 100
117
  },
118
+ "SoccerTwos.IsTraining.sum": {
119
  "value": 1.0,
120
  "min": 1.0,
121
  "max": 1.0,
122
+ "count": 100
123
+ },
124
+ "SoccerTwos.Losses.PolicyLoss.mean": {
125
+ "value": 0.0168987456165875,
126
+ "min": 0.012042445930031438,
127
+ "max": 0.02239705673904003,
128
+ "count": 46
129
+ },
130
+ "SoccerTwos.Losses.PolicyLoss.sum": {
131
+ "value": 0.0168987456165875,
132
+ "min": 0.012042445930031438,
133
+ "max": 0.02239705673904003,
134
+ "count": 46
135
+ },
136
+ "SoccerTwos.Losses.ValueLoss.mean": {
137
+ "value": 0.000625235964738143,
138
+ "min": 4.422941644103654e-06,
139
+ "max": 0.007038854543740551,
140
+ "count": 46
141
+ },
142
+ "SoccerTwos.Losses.ValueLoss.sum": {
143
+ "value": 0.000625235964738143,
144
+ "min": 4.422941644103654e-06,
145
+ "max": 0.007038854543740551,
146
+ "count": 46
147
+ },
148
+ "SoccerTwos.Losses.BaselineLoss.mean": {
149
+ "value": 0.0006257908535189927,
150
+ "min": 4.599396841816391e-06,
151
+ "max": 0.0070884921588003635,
152
+ "count": 46
153
+ },
154
+ "SoccerTwos.Losses.BaselineLoss.sum": {
155
+ "value": 0.0006257908535189927,
156
+ "min": 4.599396841816391e-06,
157
+ "max": 0.0070884921588003635,
158
+ "count": 46
159
+ },
160
+ "SoccerTwos.Policy.LearningRate.mean": {
161
+ "value": 0.0003,
162
+ "min": 0.0003,
163
+ "max": 0.0003,
164
+ "count": 46
165
+ },
166
+ "SoccerTwos.Policy.LearningRate.sum": {
167
+ "value": 0.0003,
168
+ "min": 0.0003,
169
+ "max": 0.0003,
170
+ "count": 46
171
+ },
172
+ "SoccerTwos.Policy.Epsilon.mean": {
173
+ "value": 0.20000000000000007,
174
+ "min": 0.2,
175
+ "max": 0.20000000000000007,
176
+ "count": 46
177
+ },
178
+ "SoccerTwos.Policy.Epsilon.sum": {
179
+ "value": 0.20000000000000007,
180
+ "min": 0.2,
181
+ "max": 0.20000000000000007,
182
+ "count": 46
183
+ },
184
+ "SoccerTwos.Policy.Beta.mean": {
185
+ "value": 0.005000000000000001,
186
+ "min": 0.005,
187
+ "max": 0.005000000000000001,
188
+ "count": 46
189
+ },
190
+ "SoccerTwos.Policy.Beta.sum": {
191
+ "value": 0.005000000000000001,
192
+ "min": 0.005,
193
+ "max": 0.005000000000000001,
194
+ "count": 46
195
  }
196
  },
197
  "metadata": {
198
  "timer_format_version": "0.1.0",
199
+ "start_time_seconds": "1699443423",
200
  "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
201
+ "command_line_arguments": "/usr/local/bin/mlagents-learn /content/ml-agents/config/poca/SoccerTwos.yaml --env=/content/ml-agents/training-envs-executables/linux/SoccerTwos --run-id=SoccerTwos --no-graphics --force",
202
  "mlagents_version": "1.1.0.dev0",
203
  "mlagents_envs_version": "1.1.0.dev0",
204
  "communication_protocol_version": "1.5.0",
205
  "pytorch_version": "2.1.0+cu118",
206
  "numpy_version": "1.23.5",
207
+ "end_time_seconds": "1699445605"
208
  },
209
+ "total": 2181.160465174,
210
  "count": 1,
211
+ "self": 0.42680890199972055,
212
  "children": {
213
  "run_training.setup": {
214
+ "total": 0.0681738690000202,
215
  "count": 1,
216
+ "self": 0.0681738690000202
217
  },
218
  "TrainerController.start_learning": {
219
+ "total": 2180.665482403,
220
  "count": 1,
221
+ "self": 1.482598219130523,
222
  "children": {
223
  "TrainerController._reset_env": {
224
+ "total": 4.778055871999641,
225
+ "count": 5,
226
+ "self": 4.778055871999641
227
  },
228
  "TrainerController.advance": {
229
+ "total": 2174.1929239598703,
230
+ "count": 64922,
231
+ "self": 1.6057936520119256,
232
  "children": {
233
  "env_step": {
234
+ "total": 1780.3524618128895,
235
+ "count": 64922,
236
+ "self": 1369.766810939087,
237
  "children": {
238
  "SubprocessEnvManager._take_step": {
239
+ "total": 409.6754709328279,
240
+ "count": 64922,
241
+ "self": 10.97205778098487,
242
  "children": {
243
  "TorchPolicy.evaluate": {
244
+ "total": 398.70341315184305,
245
+ "count": 129046,
246
+ "self": 398.70341315184305
247
  }
248
  }
249
  },
250
  "workers": {
251
+ "total": 0.9101799409745581,
252
+ "count": 64922,
253
  "self": 0.0,
254
  "children": {
255
  "worker_root": {
256
+ "total": 2176.474749460971,
257
+ "count": 64922,
258
  "is_parallel": true,
259
+ "self": 1019.283098851889,
260
  "children": {
261
  "run_training.setup": {
262
  "total": 0.0,
 
265
  "self": 0.0,
266
  "children": {
267
  "steps_from_proto": {
268
+ "total": 0.004532323999228538,
269
+ "count": 2,
270
  "is_parallel": true,
271
+ "self": 0.0010208789999524015,
272
  "children": {
273
  "_process_rank_one_or_two_observation": {
274
+ "total": 0.0035114449992761365,
275
+ "count": 8,
276
  "is_parallel": true,
277
+ "self": 0.0035114449992761365
278
  }
279
  }
280
  },
281
  "UnityEnvironment.step": {
282
+ "total": 0.03913147999992361,
283
  "count": 1,
284
  "is_parallel": true,
285
+ "self": 0.001174363998870831,
286
  "children": {
287
  "UnityEnvironment._generate_step_input": {
288
+ "total": 0.000787323000622564,
289
  "count": 1,
290
  "is_parallel": true,
291
+ "self": 0.000787323000622564
292
  },
293
  "communicator.exchange": {
294
+ "total": 0.03343747400049324,
295
  "count": 1,
296
  "is_parallel": true,
297
+ "self": 0.03343747400049324
298
  },
299
  "steps_from_proto": {
300
+ "total": 0.003732318999936979,
301
+ "count": 2,
302
  "is_parallel": true,
303
+ "self": 0.0007477830013158382,
304
  "children": {
305
  "_process_rank_one_or_two_observation": {
306
+ "total": 0.0029845359986211406,
307
+ "count": 8,
308
  "is_parallel": true,
309
+ "self": 0.0029845359986211406
310
  }
311
  }
312
  }
 
315
  }
316
  },
317
  "UnityEnvironment.step": {
318
+ "total": 1157.181611549082,
319
+ "count": 64921,
320
  "is_parallel": true,
321
+ "self": 75.01715701136709,
322
  "children": {
323
  "UnityEnvironment._generate_step_input": {
324
+ "total": 48.206555096093325,
325
+ "count": 64921,
326
  "is_parallel": true,
327
+ "self": 48.206555096093325
328
  },
329
  "communicator.exchange": {
330
+ "total": 806.254951789816,
331
+ "count": 64921,
332
  "is_parallel": true,
333
+ "self": 806.254951789816
334
  },
335
  "steps_from_proto": {
336
+ "total": 227.70294765180552,
337
+ "count": 129842,
338
  "is_parallel": true,
339
+ "self": 36.63705008140914,
340
  "children": {
341
  "_process_rank_one_or_two_observation": {
342
+ "total": 191.06589757039637,
343
+ "count": 519368,
344
  "is_parallel": true,
345
+ "self": 191.06589757039637
346
  }
347
  }
348
  }
349
  }
350
+ },
351
+ "steps_from_proto": {
352
+ "total": 0.010039060000053723,
353
+ "count": 8,
354
+ "is_parallel": true,
355
+ "self": 0.0018889550001404132,
356
+ "children": {
357
+ "_process_rank_one_or_two_observation": {
358
+ "total": 0.00815010499991331,
359
+ "count": 32,
360
+ "is_parallel": true,
361
+ "self": 0.00815010499991331
362
+ }
363
+ }
364
  }
365
  }
366
  }
367
  }
368
  }
369
  }
370
+ },
371
+ "trainer_advance": {
372
+ "total": 392.2346684949689,
373
+ "count": 64922,
374
+ "self": 14.265407442085234,
 
 
 
 
 
 
 
 
375
  "children": {
376
+ "process_trajectory": {
377
+ "total": 105.45076121288639,
378
+ "count": 64922,
379
+ "self": 104.89955790788736,
 
380
  "children": {
381
+ "RLTrainer._checkpoint": {
382
+ "total": 0.551203304999035,
383
+ "count": 2,
384
+ "self": 0.551203304999035
385
+ }
386
+ }
387
+ },
388
+ "_update_policy": {
389
+ "total": 272.5184998399973,
390
+ "count": 46,
391
+ "self": 164.45873168098115,
392
+ "children": {
393
+ "TorchPOCAOptimizer.update": {
394
+ "total": 108.05976815901613,
395
+ "count": 1383,
396
+ "self": 108.05976815901613
 
 
 
 
 
 
 
 
 
 
 
397
  }
398
  }
399
  }
 
401
  }
402
  }
403
  },
404
+ "trainer_threads": {
405
+ "total": 9.99999429041054e-07,
406
+ "count": 1,
407
+ "self": 9.99999429041054e-07
408
+ },
409
  "TrainerController._save_models": {
410
+ "total": 0.2119033520002631,
411
  "count": 1,
412
+ "self": 0.0022559970002475893,
413
  "children": {
414
  "RLTrainer._checkpoint": {
415
+ "total": 0.20964735500001552,
416
  "count": 1,
417
+ "self": 0.20964735500001552
418
  }
419
  }
420
  }
run_logs/training_status.json CHANGED
@@ -1,59 +1,42 @@
1
  {
2
- "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
- "steps": 49936,
6
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 12.636363636363637,
8
- "creation_time": 1698484933.1987317,
9
  "auxillary_file_paths": [
10
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
12
  },
13
  {
14
- "steps": 99960,
15
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 21.818181818181817,
17
- "creation_time": 1698485059.0796082,
18
  "auxillary_file_paths": [
19
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
21
  },
22
  {
23
- "steps": 149984,
24
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 23.727272727272727,
26
- "creation_time": 1698485184.1361842,
27
  "auxillary_file_paths": [
28
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
- ]
30
- },
31
- {
32
- "steps": 199984,
33
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": 27.363636363636363,
35
- "creation_time": 1698485308.953085,
36
- "auxillary_file_paths": [
37
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
- ]
39
- },
40
- {
41
- "steps": 200112,
42
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
- "reward": 27.363636363636363,
44
- "creation_time": 1698485309.061246,
45
- "auxillary_file_paths": [
46
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
48
  }
49
  ],
 
50
  "final_checkpoint": {
51
- "steps": 200112,
52
- "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
- "reward": 27.363636363636363,
54
- "creation_time": 1698485309.061246,
55
  "auxillary_file_paths": [
56
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]
58
  }
59
  },
 
1
  {
2
+ "SoccerTwos": {
3
  "checkpoints": [
4
  {
5
+ "steps": 499118,
6
+ "file_path": "results/SoccerTwos/SoccerTwos/SoccerTwos-499118.onnx",
7
+ "reward": 0.0,
8
+ "creation_time": 1699444487.060512,
9
  "auxillary_file_paths": [
10
+ "results/SoccerTwos/SoccerTwos/SoccerTwos-499118.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 999986,
15
+ "file_path": "results/SoccerTwos/SoccerTwos/SoccerTwos-999986.onnx",
16
+ "reward": 0.0,
17
+ "creation_time": 1699445604.2686276,
18
  "auxillary_file_paths": [
19
+ "results/SoccerTwos/SoccerTwos/SoccerTwos-999986.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 1001986,
24
+ "file_path": "results/SoccerTwos/SoccerTwos/SoccerTwos-1001986.onnx",
25
+ "reward": 0.0,
26
+ "creation_time": 1699445604.6886976,
27
  "auxillary_file_paths": [
28
+ "results/SoccerTwos/SoccerTwos/SoccerTwos-1001986.pt"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  ]
30
  }
31
  ],
32
+ "elo": 1197.1405668605373,
33
  "final_checkpoint": {
34
+ "steps": 1001986,
35
+ "file_path": "results/SoccerTwos/SoccerTwos.onnx",
36
+ "reward": 0.0,
37
+ "creation_time": 1699445604.6886976,
38
  "auxillary_file_paths": [
39
+ "results/SoccerTwos/SoccerTwos/SoccerTwos-1001986.pt"
40
  ]
41
  }
42
  },