gabrielbenabou commited on
Commit
5028df6
1 Parent(s): 9f91a2d

First Push

Browse files
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d53c31fb07fbed8774b85cfc7152ccf7507a01eb7666ad36ce6c578c9b7e1b1e
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:535bae02ab217c2ae233491f22a90f9da14257ad02f1fcf3e608ece560a1ed91
3
  size 650646
SnowballTarget/SnowballTarget-149984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02b6f032fef328725757e38a55a764acad73712c7a9f39ff9dba0d588c35188f
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f5c25dfbbb70e6caa542889a595eda772a5e0a3cdd3f154c7a49524db3acd61
3
  size 650646
SnowballTarget/SnowballTarget-149984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6043378c34ad7bfb4e2dba2b77f733699907964a94a84443577ee1c04b13da2
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48d3565c15b2677d71ef8e062e8720cce1e192655fe477aad5d9b039b13b51ce
3
  size 3850075
SnowballTarget/SnowballTarget-199984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d53c31fb07fbed8774b85cfc7152ccf7507a01eb7666ad36ce6c578c9b7e1b1e
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:535bae02ab217c2ae233491f22a90f9da14257ad02f1fcf3e608ece560a1ed91
3
  size 650646
SnowballTarget/SnowballTarget-199984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdb8be767190b567242e6c9022ceecf14039cde6cdf8d1419a832ff99f7d59aa
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7601510b05059f78af7c9fb8a3912a459a3b37f93c29f51f2c0c744d3f0a0b11
3
  size 3850075
SnowballTarget/SnowballTarget-200112.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d53c31fb07fbed8774b85cfc7152ccf7507a01eb7666ad36ce6c578c9b7e1b1e
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:535bae02ab217c2ae233491f22a90f9da14257ad02f1fcf3e608ece560a1ed91
3
  size 650646
SnowballTarget/SnowballTarget-200112.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba2ca63b9626343f889c0af01f847a514034afd9ec8fb50b4b62e38618d0221b
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3512c8123a74048a278bcef04b71d6ce7d830445b8fe96e12fd378c606ec7199
3
  size 3850075
SnowballTarget/SnowballTarget-49936.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fef76f8d697cf83e8053be1a34cd52054667b561b3586c83ac040944856dcb17
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49d6343c0efe58a6f3b3125b7a1f7c1d1f366c03870a8196808d5781b282640a
3
  size 650646
SnowballTarget/SnowballTarget-49936.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fa0c5a79420b59f1e342e279420d4d5b1265410c45772f0b8a377727c260318
3
  size 3850000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5c41d496d0576c41ce13eca98c6b21904130120c235f8bf3696a576189b99d2
3
  size 3850000
SnowballTarget/SnowballTarget-99960.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2c3358db10d2d904b4d6222123870bbe6d91f09ba1a1a56aaa46ff3e5e495a2
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c0ff36ced7856ea1f3a67511e453396511d372ab2b3d7f5c3100f20920b2a2e
3
  size 650646
SnowballTarget/SnowballTarget-99960.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91a304d654701e939c8a701ecc0fa8a3ed2577d6a5c84e0e039101eed7763d75
3
  size 3850000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be307f36c84a51cb20ab5e013203bfd7fa21d24aa113c066df50a7805dcd1e77
3
  size 3850000
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9cfa14e1509fd2c973ef5b0d53440fdc52611529b4a9dafcb8158d66689e6808
3
  size 3849250
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3317d42e91e3d2b1dd6e911a70d25514ca2824193f6609d48cb1f491285920be
3
  size 3849250
SnowballTarget/events.out.tfevents.1708079523.e4fd9420214f.1594.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d52bbcc01d37f5fea0d78b318fc624df6ae1aba36dc258118052348a31d1c298
3
+ size 28879
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -70,7 +70,7 @@ checkpoint_settings:
70
  initialize_from: null
71
  load_model: false
72
  resume: false
73
- force: false
74
  train_model: false
75
  inference: false
76
  results_dir: results
 
70
  initialize_from: null
71
  load_model: false
72
  resume: false
73
+ force: true
74
  train_model: false
75
  inference: false
76
  results_dir: results
run_logs/Player-0.log CHANGED
@@ -1,12 +1,9 @@
1
- Mono path[0] = '/workspaces/reinforcement-learning-course/course/unity5/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/Managed'
2
- Mono config path = '/workspaces/reinforcement-learning-course/course/unity5/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face
6
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face/SnowballTarget
7
- Unable to load player prefs
8
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
- [Subsystems] Discovering subsystems at path /workspaces/reinforcement-learning-course/course/unity5/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
10
  Forcing GfxDevice: Null
11
  GfxDevice: creating device client; threaded=0; jobified=0
12
  NullGfxDevice:
@@ -34,7 +31,7 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.042 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -45,7 +42,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.420478 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 16 worker threads for Enlighten.
@@ -98,12 +95,12 @@ Memory Statistics:
98
  Current Block Size 256.0 KB
99
  Peak Allocated Bytes 3.6 KB
100
  Overflow Count 0
101
- [ALLOC_TEMP_Job.Worker 14]
102
  Initial Block Size 256.0 KB
103
  Current Block Size 256.0 KB
104
  Peak Allocated Bytes 3.6 KB
105
  Overflow Count 0
106
- [ALLOC_TEMP_Job.Worker 17]
107
  Initial Block Size 256.0 KB
108
  Current Block Size 256.0 KB
109
  Peak Allocated Bytes 3.6 KB
@@ -193,12 +190,12 @@ Memory Statistics:
193
  Current Block Size 256.0 KB
194
  Peak Allocated Bytes 3.6 KB
195
  Overflow Count 0
196
- [ALLOC_TEMP_Job.Worker 26]
197
  Initial Block Size 256.0 KB
198
  Current Block Size 256.0 KB
199
  Peak Allocated Bytes 3.6 KB
200
  Overflow Count 0
201
- [ALLOC_TEMP_Job.Worker 21]
202
  Initial Block Size 256.0 KB
203
  Current Block Size 256.0 KB
204
  Peak Allocated Bytes 3.6 KB
@@ -208,7 +205,7 @@ Memory Statistics:
208
  Current Block Size 256.0 KB
209
  Peak Allocated Bytes 3.6 KB
210
  Overflow Count 0
211
- [ALLOC_TEMP_Background Job.Worker 14]
212
  Initial Block Size 32.0 KB
213
  Current Block Size 32.0 KB
214
  Peak Allocated Bytes 0 B
@@ -218,7 +215,7 @@ Memory Statistics:
218
  Current Block Size 256.0 KB
219
  Peak Allocated Bytes 3.6 KB
220
  Overflow Count 0
221
- [ALLOC_TEMP_Background Job.Worker 9]
222
  Initial Block Size 32.0 KB
223
  Current Block Size 32.0 KB
224
  Peak Allocated Bytes 0 B
@@ -293,22 +290,22 @@ Memory Statistics:
293
  Current Block Size 256.0 KB
294
  Peak Allocated Bytes 3.6 KB
295
  Overflow Count 0
296
- [ALLOC_TEMP_Job.Worker 15]
297
  Initial Block Size 256.0 KB
298
  Current Block Size 256.0 KB
299
  Peak Allocated Bytes 3.6 KB
300
  Overflow Count 0
301
- [ALLOC_TEMP_Job.Worker 24]
302
  Initial Block Size 256.0 KB
303
  Current Block Size 256.0 KB
304
  Peak Allocated Bytes 3.6 KB
305
  Overflow Count 0
306
- [ALLOC_TEMP_Job.Worker 19]
307
  Initial Block Size 256.0 KB
308
  Current Block Size 256.0 KB
309
  Peak Allocated Bytes 3.6 KB
310
  Overflow Count 0
311
- [ALLOC_TEMP_Job.Worker 12]
312
  Initial Block Size 256.0 KB
313
  Current Block Size 256.0 KB
314
  Peak Allocated Bytes 3.6 KB
@@ -316,10 +313,10 @@ Memory Statistics:
316
  [ALLOC_TEMP_Loading.AsyncRead]
317
  Initial Block Size 64.0 KB
318
  Current Block Size 64.0 KB
319
- Peak Allocated Bytes 292 B
320
  Overflow Count 0
321
  [ALLOC_DEFAULT] Dual Thread Allocator
322
- Peak main deferred allocation count 41
323
  [ALLOC_BUCKET]
324
  Large Block size 4.0 MB
325
  Used Block count 1
@@ -334,7 +331,7 @@ Memory Statistics:
334
  Peak usage frame count: [16.0 MB-32.0 MB]: 5461 frames
335
  Requested Block Size 16.0 MB
336
  Peak Block count 1
337
- Peak Allocated memory 17.6 MB
338
  Peak Large allocation bytes 16.0 MB
339
  [ALLOC_TEMP_JOB_1_FRAME]
340
  Initial Block Size 2.0 MB
@@ -363,10 +360,10 @@ Memory Statistics:
363
  Used Block count 1
364
  Peak Allocated bytes 1.0 MB
365
  [ALLOC_GFX_MAIN]
366
- Peak usage frame count: [32.0 KB-64.0 KB]: 4171 frames, [64.0 KB-128.0 KB]: 1290 frames
367
  Requested Block Size 16.0 MB
368
  Peak Block count 1
369
- Peak Allocated memory 67.3 KB
370
  Peak Large allocation bytes 0 B
371
  [ALLOC_GFX_THREAD]
372
  Peak usage frame count: [32.0 KB-64.0 KB]: 5461 frames
 
1
+ Mono path[0] = '/workspaces/reinforcement-learning-course/course/unit5/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/Managed'
2
+ Mono config path = '/workspaces/reinforcement-learning-course/course/unit5/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
5
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
+ [Subsystems] Discovering subsystems at path /workspaces/reinforcement-learning-course/course/unit5/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
8
  GfxDevice: creating device client; threaded=0; jobified=0
9
  NullGfxDevice:
 
31
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
  FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
+ - Completed reload, in 0.039 seconds
35
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
42
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
43
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
44
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
45
+ UnloadTime: 0.408071 ms
46
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
47
  requesting resize 84 x 84
48
  Setting up 16 worker threads for Enlighten.
 
95
  Current Block Size 256.0 KB
96
  Peak Allocated Bytes 3.6 KB
97
  Overflow Count 0
98
+ [ALLOC_TEMP_Job.Worker 17]
99
  Initial Block Size 256.0 KB
100
  Current Block Size 256.0 KB
101
  Peak Allocated Bytes 3.6 KB
102
  Overflow Count 0
103
+ [ALLOC_TEMP_Job.Worker 14]
104
  Initial Block Size 256.0 KB
105
  Current Block Size 256.0 KB
106
  Peak Allocated Bytes 3.6 KB
 
190
  Current Block Size 256.0 KB
191
  Peak Allocated Bytes 3.6 KB
192
  Overflow Count 0
193
+ [ALLOC_TEMP_Job.Worker 21]
194
  Initial Block Size 256.0 KB
195
  Current Block Size 256.0 KB
196
  Peak Allocated Bytes 3.6 KB
197
  Overflow Count 0
198
+ [ALLOC_TEMP_Job.Worker 26]
199
  Initial Block Size 256.0 KB
200
  Current Block Size 256.0 KB
201
  Peak Allocated Bytes 3.6 KB
 
205
  Current Block Size 256.0 KB
206
  Peak Allocated Bytes 3.6 KB
207
  Overflow Count 0
208
+ [ALLOC_TEMP_Background Job.Worker 9]
209
  Initial Block Size 32.0 KB
210
  Current Block Size 32.0 KB
211
  Peak Allocated Bytes 0 B
 
215
  Current Block Size 256.0 KB
216
  Peak Allocated Bytes 3.6 KB
217
  Overflow Count 0
218
+ [ALLOC_TEMP_Background Job.Worker 14]
219
  Initial Block Size 32.0 KB
220
  Current Block Size 32.0 KB
221
  Peak Allocated Bytes 0 B
 
290
  Current Block Size 256.0 KB
291
  Peak Allocated Bytes 3.6 KB
292
  Overflow Count 0
293
+ [ALLOC_TEMP_Job.Worker 24]
294
  Initial Block Size 256.0 KB
295
  Current Block Size 256.0 KB
296
  Peak Allocated Bytes 3.6 KB
297
  Overflow Count 0
298
+ [ALLOC_TEMP_Job.Worker 15]
299
  Initial Block Size 256.0 KB
300
  Current Block Size 256.0 KB
301
  Peak Allocated Bytes 3.6 KB
302
  Overflow Count 0
303
+ [ALLOC_TEMP_Job.Worker 12]
304
  Initial Block Size 256.0 KB
305
  Current Block Size 256.0 KB
306
  Peak Allocated Bytes 3.6 KB
307
  Overflow Count 0
308
+ [ALLOC_TEMP_Job.Worker 19]
309
  Initial Block Size 256.0 KB
310
  Current Block Size 256.0 KB
311
  Peak Allocated Bytes 3.6 KB
 
313
  [ALLOC_TEMP_Loading.AsyncRead]
314
  Initial Block Size 64.0 KB
315
  Current Block Size 64.0 KB
316
+ Peak Allocated Bytes 290 B
317
  Overflow Count 0
318
  [ALLOC_DEFAULT] Dual Thread Allocator
319
+ Peak main deferred allocation count 40
320
  [ALLOC_BUCKET]
321
  Large Block size 4.0 MB
322
  Used Block count 1
 
331
  Peak usage frame count: [16.0 MB-32.0 MB]: 5461 frames
332
  Requested Block Size 16.0 MB
333
  Peak Block count 1
334
+ Peak Allocated memory 17.5 MB
335
  Peak Large allocation bytes 16.0 MB
336
  [ALLOC_TEMP_JOB_1_FRAME]
337
  Initial Block Size 2.0 MB
 
360
  Used Block count 1
361
  Peak Allocated bytes 1.0 MB
362
  [ALLOC_GFX_MAIN]
363
+ Peak usage frame count: [32.0 KB-64.0 KB]: 4530 frames, [64.0 KB-128.0 KB]: 931 frames
364
  Requested Block Size 16.0 MB
365
  Peak Block count 1
366
+ Peak Allocated memory 66.8 KB
367
  Peak Large allocation bytes 0 B
368
  [ALLOC_GFX_THREAD]
369
  Peak usage frame count: [32.0 KB-64.0 KB]: 5461 frames
run_logs/timers.json CHANGED
@@ -2,15 +2,15 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.9938374161720276,
6
- "min": 0.9938374161720276,
7
- "max": 2.867473840713501,
8
  "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 9521.9560546875,
12
- "min": 9521.9560546875,
13
- "max": 29460.42578125,
14
  "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
@@ -26,15 +26,15 @@
26
  "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 11.73255443572998,
30
- "min": 0.5038750767707825,
31
- "max": 11.73255443572998,
32
  "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2287.84814453125,
36
- "min": 97.75177001953125,
37
- "max": 2379.9619140625,
38
  "count": 20
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
@@ -49,34 +49,58 @@
49
  "max": 10945.0,
50
  "count": 20
51
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.07142672198881651,
54
- "min": 0.06495762881868537,
55
- "max": 0.07605622582309925,
56
  "count": 20
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.28570688795526605,
60
- "min": 0.2598305152747415,
61
- "max": 0.3785982645760455,
62
  "count": 20
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.19691719326610657,
66
- "min": 0.12123059188974472,
67
- "max": 0.2905612743514426,
68
  "count": 20
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.7876687730644263,
72
- "min": 0.4849223675589789,
73
- "max": 1.4071116996746436,
74
  "count": 20
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
  "value": 8.082097306000005e-06,
78
  "min": 8.082097306000005e-06,
79
- "max": 0.000291882002706,
80
  "count": 20
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
@@ -88,7 +112,7 @@
88
  "SnowballTarget.Policy.Epsilon.mean": {
89
  "value": 0.10269400000000001,
90
  "min": 0.10269400000000001,
91
- "max": 0.19729400000000002,
92
  "count": 20
93
  },
94
  "SnowballTarget.Policy.Epsilon.sum": {
@@ -100,7 +124,7 @@
100
  "SnowballTarget.Policy.Beta.mean": {
101
  "value": 0.0001444306000000001,
102
  "min": 0.0001444306000000001,
103
- "max": 0.0048649706,
104
  "count": 20
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
@@ -109,30 +133,6 @@
109
  "max": 0.023089828,
110
  "count": 20
111
  },
112
- "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 23.15909090909091,
114
- "min": 3.5,
115
- "max": 23.527272727272727,
116
- "count": 20
117
- },
118
- "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1019.0,
120
- "min": 154.0,
121
- "max": 1294.0,
122
- "count": 20
123
- },
124
- "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 23.15909090909091,
126
- "min": 3.5,
127
- "max": 23.527272727272727,
128
- "count": 20
129
- },
130
- "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1019.0,
132
- "min": 154.0,
133
- "max": 1294.0,
134
- "count": 20
135
- },
136
  "SnowballTarget.IsTraining.mean": {
137
  "value": 1.0,
138
  "min": 1.0,
@@ -148,67 +148,67 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1708005961",
152
  "python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
153
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.1.0.dev0",
155
  "mlagents_envs_version": "1.1.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.2.0+cu121",
158
- "numpy_version": "1.23.5",
159
- "end_time_seconds": "1708006338"
160
  },
161
- "total": 376.8408562250006,
162
  "count": 1,
163
- "self": 0.22000378999655368,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.01569223200203851,
167
  "count": 1,
168
- "self": 0.01569223200203851
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 376.605160203002,
172
  "count": 1,
173
- "self": 0.3858108900967636,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 1.2901278770004865,
177
  "count": 1,
178
- "self": 1.2901278770004865
179
  },
180
  "TrainerController.advance": {
181
- "total": 374.87211922190545,
182
- "count": 18202,
183
- "self": 0.1873388351741596,
184
  "children": {
185
  "env_step": {
186
- "total": 374.6847803867313,
187
- "count": 18202,
188
- "self": 216.90356242378766,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 157.56626354416585,
192
- "count": 18202,
193
- "self": 1.112029014449945,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 156.4542345297159,
197
- "count": 18202,
198
- "self": 156.4542345297159
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.2149544187777792,
204
- "count": 18202,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 375.8842824683743,
209
- "count": 18202,
210
  "is_parallel": true,
211
- "self": 215.02137959124957,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.0013307830013218336,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.00047157200242509134,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0008592109988967422,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0008592109988967422
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.020179667000775225,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.00019189000522601418,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00014331700003822334,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00014331700003822334
244
  },
245
  "communicator.exchange": {
246
- "total": 0.01915219899819931,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.01915219899819931
250
  },
251
  "steps_from_proto": {
252
- "total": 0.0006922609973116778,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.00017056699653039686,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.0005216940007812809,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.0005216940007812809
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 160.86290287712472,
271
- "count": 18201,
272
  "is_parallel": true,
273
- "self": 3.50999104643779,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 2.147183898217918,
277
- "count": 18201,
278
  "is_parallel": true,
279
- "self": 2.147183898217918
280
  },
281
  "communicator.exchange": {
282
- "total": 145.64208123623393,
283
- "count": 18201,
284
  "is_parallel": true,
285
- "self": 145.64208123623393
286
  },
287
  "steps_from_proto": {
288
- "total": 9.563646696235082,
289
- "count": 18201,
290
  "is_parallel": true,
291
- "self": 2.235202875788673,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 7.3284438204464095,
295
- "count": 182010,
296
  "is_parallel": true,
297
- "self": 7.3284438204464095
298
  }
299
  }
300
  }
@@ -309,9 +309,9 @@
309
  }
310
  },
311
  "trainer_threads": {
312
- "total": 0.00010352599929319695,
313
  "count": 1,
314
- "self": 0.00010352599929319695,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
@@ -320,36 +320,36 @@
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
- "total": 373.14890932762137,
324
- "count": 332617,
325
  "is_parallel": true,
326
- "self": 5.223721549289621,
327
  "children": {
328
  "process_trajectory": {
329
- "total": 209.01437588933914,
330
- "count": 332617,
331
  "is_parallel": true,
332
- "self": 208.58555328633884,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
- "total": 0.42882260300029884,
336
  "count": 4,
337
  "is_parallel": true,
338
- "self": 0.42882260300029884
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
- "total": 158.9108118889926,
344
  "count": 90,
345
  "is_parallel": true,
346
- "self": 23.660605850047432,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
- "total": 135.25020603894518,
350
  "count": 4587,
351
  "is_parallel": true,
352
- "self": 135.25020603894518
353
  }
354
  }
355
  }
@@ -360,14 +360,14 @@
360
  }
361
  },
362
  "TrainerController._save_models": {
363
- "total": 0.0569986880000215,
364
  "count": 1,
365
- "self": 0.0006153909998829477,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
- "total": 0.056383297000138555,
369
  "count": 1,
370
- "self": 0.056383297000138555
371
  }
372
  }
373
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.9192315340042114,
6
+ "min": 0.9192315340042114,
7
+ "max": 2.8722503185272217,
8
  "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 8797.0458984375,
12
+ "min": 8797.0458984375,
13
+ "max": 29477.904296875,
14
  "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
 
26
  "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 12.675583839416504,
30
+ "min": 0.1918199062347412,
31
+ "max": 12.675583839416504,
32
  "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2471.73876953125,
36
+ "min": 37.21306228637695,
37
+ "max": 2572.742431640625,
38
  "count": 20
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
 
49
  "max": 10945.0,
50
  "count": 20
51
  },
52
+ "SnowballTarget.Environment.CumulativeReward.mean": {
53
+ "value": 25.09090909090909,
54
+ "min": 3.159090909090909,
55
+ "max": 25.09090909090909,
56
+ "count": 20
57
+ },
58
+ "SnowballTarget.Environment.CumulativeReward.sum": {
59
+ "value": 1104.0,
60
+ "min": 139.0,
61
+ "max": 1367.0,
62
+ "count": 20
63
+ },
64
+ "SnowballTarget.Policy.ExtrinsicReward.mean": {
65
+ "value": 25.09090909090909,
66
+ "min": 3.159090909090909,
67
+ "max": 25.09090909090909,
68
+ "count": 20
69
+ },
70
+ "SnowballTarget.Policy.ExtrinsicReward.sum": {
71
+ "value": 1104.0,
72
+ "min": 139.0,
73
+ "max": 1367.0,
74
+ "count": 20
75
+ },
76
  "SnowballTarget.Losses.PolicyLoss.mean": {
77
+ "value": 0.06375053759763308,
78
+ "min": 0.06285092248184738,
79
+ "max": 0.07566402355555957,
80
  "count": 20
81
  },
82
  "SnowballTarget.Losses.PolicyLoss.sum": {
83
+ "value": 0.2550021503905323,
84
+ "min": 0.2550021503905323,
85
+ "max": 0.37832011777779784,
86
  "count": 20
87
  },
88
  "SnowballTarget.Losses.ValueLoss.mean": {
89
+ "value": 0.20342359691858292,
90
+ "min": 0.10652379825820818,
91
+ "max": 0.2963242155079748,
92
  "count": 20
93
  },
94
  "SnowballTarget.Losses.ValueLoss.sum": {
95
+ "value": 0.8136943876743317,
96
+ "min": 0.4260951930328327,
97
+ "max": 1.481621077539874,
98
  "count": 20
99
  },
100
  "SnowballTarget.Policy.LearningRate.mean": {
101
  "value": 8.082097306000005e-06,
102
  "min": 8.082097306000005e-06,
103
+ "max": 0.00029184900271699997,
104
  "count": 20
105
  },
106
  "SnowballTarget.Policy.LearningRate.sum": {
 
112
  "SnowballTarget.Policy.Epsilon.mean": {
113
  "value": 0.10269400000000001,
114
  "min": 0.10269400000000001,
115
+ "max": 0.19728300000000004,
116
  "count": 20
117
  },
118
  "SnowballTarget.Policy.Epsilon.sum": {
 
124
  "SnowballTarget.Policy.Beta.mean": {
125
  "value": 0.0001444306000000001,
126
  "min": 0.0001444306000000001,
127
+ "max": 0.0048644217,
128
  "count": 20
129
  },
130
  "SnowballTarget.Policy.Beta.sum": {
 
133
  "max": 0.023089828,
134
  "count": 20
135
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
136
  "SnowballTarget.IsTraining.mean": {
137
  "value": 1.0,
138
  "min": 1.0,
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1708079522",
152
  "python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
153
+ "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics --force",
154
  "mlagents_version": "1.1.0.dev0",
155
  "mlagents_envs_version": "1.1.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.2.0+cu121",
158
+ "numpy_version": "1.21.2",
159
+ "end_time_seconds": "1708079916"
160
  },
161
+ "total": 393.62493889100006,
162
  "count": 1,
163
+ "self": 0.2703265179952723,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.01448039500246523,
167
  "count": 1,
168
+ "self": 0.01448039500246523
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 393.3401319780023,
172
  "count": 1,
173
+ "self": 0.43650310122029623,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 1.105587543999718,
177
  "count": 1,
178
+ "self": 1.105587543999718
179
  },
180
  "TrainerController.advance": {
181
+ "total": 391.7385997017773,
182
+ "count": 18201,
183
+ "self": 0.20346960727329133,
184
  "children": {
185
  "env_step": {
186
+ "total": 391.535130094504,
187
+ "count": 18201,
188
+ "self": 244.68716851486533,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 146.6104092521491,
192
+ "count": 18201,
193
+ "self": 1.1776646150974557,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 145.43274463705166,
197
+ "count": 18201,
198
+ "self": 145.43274463705166
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 0.23755232748953858,
204
+ "count": 18201,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 392.62108338120015,
209
+ "count": 18201,
210
  "is_parallel": true,
211
+ "self": 205.60812447789795,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.0012301889946684241,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.000406236998969689,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.0008239519956987351,
227
  "count": 10,
228
  "is_parallel": true,
229
+ "self": 0.0008239519956987351
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.023591150995343924,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.0002002689871005714,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.0005734540027333423,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.0005734540027333423
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.022032102999219205,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.022032102999219205
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.0007853250062908046,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.00019681201229104772,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0005885129939997569,
259
  "count": 10,
260
  "is_parallel": true,
261
+ "self": 0.0005885129939997569
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 187.0129589033022,
271
+ "count": 18200,
272
  "is_parallel": true,
273
+ "self": 3.8799755034633563,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 2.39522441214649,
277
+ "count": 18200,
278
  "is_parallel": true,
279
+ "self": 2.39522441214649
280
  },
281
  "communicator.exchange": {
282
+ "total": 168.77685234500677,
283
+ "count": 18200,
284
  "is_parallel": true,
285
+ "self": 168.77685234500677
286
  },
287
  "steps_from_proto": {
288
+ "total": 11.960906642685586,
289
+ "count": 18200,
290
  "is_parallel": true,
291
+ "self": 2.6766800949480967,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 9.28422654773749,
295
+ "count": 182000,
296
  "is_parallel": true,
297
+ "self": 9.28422654773749
298
  }
299
  }
300
  }
 
309
  }
310
  },
311
  "trainer_threads": {
312
+ "total": 8.46290058689192e-05,
313
  "count": 1,
314
+ "self": 8.46290058689192e-05,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
 
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
+ "total": 389.4141271024273,
324
+ "count": 344416,
325
  "is_parallel": true,
326
+ "self": 6.619416726120107,
327
  "children": {
328
  "process_trajectory": {
329
+ "total": 216.45801907726127,
330
+ "count": 344416,
331
  "is_parallel": true,
332
+ "self": 216.09312064325786,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
+ "total": 0.3648984340034076,
336
  "count": 4,
337
  "is_parallel": true,
338
+ "self": 0.3648984340034076
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
+ "total": 166.33669129904592,
344
  "count": 90,
345
  "is_parallel": true,
346
+ "self": 31.018204163425253,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
+ "total": 135.31848713562067,
350
  "count": 4587,
351
  "is_parallel": true,
352
+ "self": 135.31848713562067
353
  }
354
  }
355
  }
 
360
  }
361
  },
362
  "TrainerController._save_models": {
363
+ "total": 0.059357001999160275,
364
  "count": 1,
365
+ "self": 0.0006415859970729798,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
+ "total": 0.058715416002087295,
369
  "count": 1,
370
+ "self": 0.058715416002087295
371
  }
372
  }
373
  }
run_logs/training_status.json CHANGED
@@ -4,8 +4,8 @@
4
  {
5
  "steps": 49936,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 13.909090909090908,
8
- "creation_time": 1708006054.7001042,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
@@ -13,8 +13,8 @@
13
  {
14
  "steps": 99960,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 20.181818181818183,
17
- "creation_time": 1708006150.1190605,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
@@ -22,8 +22,8 @@
22
  {
23
  "steps": 149984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 22.454545454545453,
26
- "creation_time": 1708006244.0212195,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
@@ -31,8 +31,8 @@
31
  {
32
  "steps": 199984,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": 23.818181818181817,
35
- "creation_time": 1708006337.885693,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
@@ -40,8 +40,8 @@
40
  {
41
  "steps": 200112,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
- "reward": 23.818181818181817,
44
- "creation_time": 1708006337.9675162,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
@@ -50,8 +50,8 @@
50
  "final_checkpoint": {
51
  "steps": 200112,
52
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
- "reward": 23.818181818181817,
54
- "creation_time": 1708006337.9675162,
55
  "auxillary_file_paths": [
56
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]
 
4
  {
5
  "steps": 49936,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
+ "reward": 13.181818181818182,
8
+ "creation_time": 1708079619.5195909,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
 
13
  {
14
  "steps": 99960,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
+ "reward": 21.363636363636363,
17
+ "creation_time": 1708079715.7734494,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
 
22
  {
23
  "steps": 149984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
+ "reward": 24.636363636363637,
26
+ "creation_time": 1708079815.9141054,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
 
31
  {
32
  "steps": 199984,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
+ "reward": 25.454545454545453,
35
+ "creation_time": 1708079916.1908214,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
 
40
  {
41
  "steps": 200112,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
+ "reward": 25.454545454545453,
44
+ "creation_time": 1708079916.2722778,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
 
50
  "final_checkpoint": {
51
  "steps": 200112,
52
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
+ "reward": 25.454545454545453,
54
+ "creation_time": 1708079916.2722778,
55
  "auxillary_file_paths": [
56
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]