bobobert4 commited on
Commit
9e832ad
1 Parent(s): e96e6c5

Curiosity reward signal enabled

Browse files
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e442260048cb1fd0580820e8e71b6cd24da625735e831be978b58d9b0be3e9f0
3
  size 260484
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64c7a453877b3eb4c12f616844f8e08e5e8226b2bfc1cde00e88151d80b1106f
3
  size 260484
SnowballTarget/SnowballTarget-305800.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d17d39a84e454db3efe81d4c32ada1b2c401d78689c44c7d7d90f12fb8e6c95
3
+ size 260484
SnowballTarget/SnowballTarget-305800.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3edbfdcc94fd4cb187fe120b83176fde9ef55cf0220ad927b35d1da7c53fabb
3
+ size 1180480
SnowballTarget/SnowballTarget-319960.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d039d15074c8128323e918f4fbff0eb3c1671f6887e512c24058c16943646dc
3
+ size 260484
SnowballTarget/SnowballTarget-319960.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63079e6bb0afda5f0a2a5f2375b8bd979ba0fc87c01033615d2e0fdaafb4e1f
3
+ size 1588526
SnowballTarget/SnowballTarget-329992.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e44085ac39a3f14642bea44edcf63f33a494837312ae29c69679de02c9ac35c7
3
+ size 260484
SnowballTarget/SnowballTarget-329992.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8763104429c14d6931fbe1577b6b6299f01d8d795f3c2757eb9b34558be9fcf0
3
+ size 1588526
SnowballTarget/SnowballTarget-339952.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91556e78bcc74f88dec1ca15848a501005b8db8c51cf923e32371968f384d3c1
3
+ size 260484
SnowballTarget/SnowballTarget-339952.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57907b60675be8ad7daf0d3d611d15803d04bb673adca06b2947ef1c0a441bad
3
+ size 1588526
SnowballTarget/SnowballTarget-349992.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e070605cad6d71c2108a2aa88fa48af92b9cf61d9c68a4be4d84774a8fe05350
3
+ size 260484
SnowballTarget/SnowballTarget-349992.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b3ce24e7d49cd2fa8e03cf5fd159f0d86ce39c84be825a8d9800233c02c2004
3
+ size 1588526
SnowballTarget/SnowballTarget-359944.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d42f97a0cc2bbad5d66ad40f7512057f7214ac338fb4fcf2e6809c96a6eb8db
3
+ size 260484
SnowballTarget/SnowballTarget-359944.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b50a7df71542a34e674dbfbf844d21e93ec618984bb61cd7a09ba066b513856
3
+ size 1588526
SnowballTarget/SnowballTarget-369984.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45dfe1d9dd1523caa78c61cda134b7fe06a0f350dfdc11ec289d15cbb54f7dd2
3
+ size 260484
SnowballTarget/SnowballTarget-369984.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11a1e94523bcfd2c00828085ffa39f6b9a48ed46dc6725842aecf1161681576a
3
+ size 1588526
SnowballTarget/SnowballTarget-379936.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60ee4b2cb2da3b2df717a7cbd05c64608e149f133cd01cf7ec8a67ac885b20b0
3
+ size 260484
SnowballTarget/SnowballTarget-379936.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffa8a61cdbef68f9ac33f29be9e747588418d028181b01be315124c045ebbe5c
3
+ size 1588526
SnowballTarget/SnowballTarget-382800.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8a54147b2b41e80a384c1054be64832c6ac3ed9f3f3d734c0bffec933e17ed2
3
+ size 260484
SnowballTarget/SnowballTarget-382800.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adfe362c5653a6f3eef7eb245392dc61c8b5bc060fa2ce0956e99028d458074a
3
+ size 990656
SnowballTarget/SnowballTarget-389976.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34004c5be2562e958bdcd9f205b0df0f9ea8902dcdbcf0bc3e6aec2c2eaeae09
3
+ size 260484
SnowballTarget/SnowballTarget-389976.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:219834d71bef8d1a7aeee375c46228b8f4e1d827f1dbe8bd099ee4ca5f75d6f6
3
+ size 1588526
SnowballTarget/SnowballTarget-399928.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64c7a453877b3eb4c12f616844f8e08e5e8226b2bfc1cde00e88151d80b1106f
3
+ size 260484
SnowballTarget/SnowballTarget-399928.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db4e2b3ee72028b313753bf3440a293abf68db625940e292d5208927e9a36d56
3
+ size 1588526
SnowballTarget/SnowballTarget-400120.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64c7a453877b3eb4c12f616844f8e08e5e8226b2bfc1cde00e88151d80b1106f
3
+ size 260484
SnowballTarget/SnowballTarget-400120.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f98be8755f9487cc04d3e5776ab35a95dbbda554b0b451627ec7d219a07d24c6
3
+ size 1588526
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe8ac531dd03c6c9329ff7a6835994a182545cd977e846220e4a347e319ddac0
3
- size 784804
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f98be8755f9487cc04d3e5776ab35a95dbbda554b0b451627ec7d219a07d24c6
3
+ size 1588526
SnowballTarget/events.out.tfevents.1689921906.4ba227ddcbc5.32434.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0407594d9f949af3fd7d74d80eda5c319b759db85d51ef22c0a2bd492f2dbc5b
3
+ size 51619
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.25, "lambd": 0.95, "num_epoch": 5, "shared_critic": true, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 10000, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 96, "summary_freq": 5000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.25, "lambd": 0.95, "num_epoch": 5, "shared_critic": true, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 10000, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 0.9, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "curiosity": {"gamma": 0.99, "strength": 0.2, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0005, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 400000, "time_horizon": 96, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTargetCurious", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -26,7 +26,7 @@ behaviors:
26
  reward_signals:
27
  extrinsic:
28
  gamma: 0.99
29
- strength: 1.0
30
  network_settings:
31
  normalize: false
32
  hidden_units: 128
@@ -35,12 +35,25 @@ behaviors:
35
  memory: null
36
  goal_conditioning_type: hyper
37
  deterministic: false
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  init_path: null
39
  keep_checkpoints: 10
40
  even_checkpoints: false
41
- max_steps: 200000
42
  time_horizon: 96
43
- summary_freq: 5000
44
  threaded: true
45
  self_play: null
46
  behavioral_cloning: null
@@ -64,11 +77,11 @@ engine_settings:
64
  no_graphics: true
65
  environment_parameters: null
66
  checkpoint_settings:
67
- run_id: SnowballTarget1
68
  initialize_from: null
69
  load_model: false
70
  resume: false
71
- force: false
72
  train_model: false
73
  inference: false
74
  results_dir: results
 
26
  reward_signals:
27
  extrinsic:
28
  gamma: 0.99
29
+ strength: 0.9
30
  network_settings:
31
  normalize: false
32
  hidden_units: 128
 
35
  memory: null
36
  goal_conditioning_type: hyper
37
  deterministic: false
38
+ curiosity:
39
+ gamma: 0.99
40
+ strength: 0.2
41
+ network_settings:
42
+ normalize: false
43
+ hidden_units: 128
44
+ num_layers: 2
45
+ vis_encode_type: simple
46
+ memory: null
47
+ goal_conditioning_type: hyper
48
+ deterministic: false
49
+ learning_rate: 0.0005
50
+ encoding_size: null
51
  init_path: null
52
  keep_checkpoints: 10
53
  even_checkpoints: false
54
+ max_steps: 400000
55
  time_horizon: 96
56
+ summary_freq: 10000
57
  threaded: true
58
  self_play: null
59
  behavioral_cloning: null
 
77
  no_graphics: true
78
  environment_parameters: null
79
  checkpoint_settings:
80
+ run_id: SnowballTargetCurious
81
  initialize_from: null
82
  load_model: false
83
  resume: false
84
+ force: true
85
  train_model: false
86
  inference: false
87
  results_dir: results
run_logs/Player-0.log CHANGED
@@ -2,9 +2,6 @@ Mono path[0] = '/content/ml-agents/training-envs-executables/linux/SnowballTarge
2
  Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face
6
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face/SnowballTarget
7
- Unable to load player prefs
8
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
  [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
10
  Forcing GfxDevice: Null
@@ -34,7 +31,7 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.082 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -45,7 +42,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.673895 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
@@ -53,7 +50,7 @@ Memory Statistics:
53
  [ALLOC_TEMP_TLS] TLS Allocator
54
  StackAllocators :
55
  [ALLOC_TEMP_MAIN]
56
- Peak usage frame count: [8.0 KB-16.0 KB]: 5374 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
57
  Initial Block Size 4.0 MB
58
  Current Block Size 4.0 MB
59
  Peak Allocated Bytes 2.0 MB
@@ -169,22 +166,22 @@ Memory Statistics:
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
- Peak main deferred allocation count 45
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
- Peak usage frame count: [4.0 MB-8.0 MB]: 5466 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
- Peak Allocated memory 6.7 MB
182
  Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
- Peak usage frame count: [16.0 MB-32.0 MB]: 5466 frames
185
  Requested Block Size 16.0 MB
186
  Peak Block count 1
187
- Peak Allocated memory 17.8 MB
188
  Peak Large allocation bytes 16.0 MB
189
  [ALLOC_TEMP_JOB_1_FRAME]
190
  Initial Block Size 2.0 MB
@@ -213,13 +210,13 @@ Memory Statistics:
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
- Peak usage frame count: [32.0 KB-64.0 KB]: 4365 frames, [64.0 KB-128.0 KB]: 1101 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
- Peak Allocated memory 67.3 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
- Peak usage frame count: [32.0 KB-64.0 KB]: 5466 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 1
225
  Peak Allocated memory 39.6 KB
@@ -231,13 +228,13 @@ Memory Statistics:
231
  Used Block count 1
232
  Peak Allocated bytes 1.0 MB
233
  [ALLOC_CACHEOBJECTS_MAIN]
234
- Peak usage frame count: [0.5 MB-1.0 MB]: 5466 frames
235
  Requested Block Size 4.0 MB
236
  Peak Block count 1
237
  Peak Allocated memory 0.6 MB
238
  Peak Large allocation bytes 0 B
239
  [ALLOC_CACHEOBJECTS_THREAD]
240
- Peak usage frame count: [0.5 MB-1.0 MB]: 5465 frames, [2.0 MB-4.0 MB]: 1 frames
241
  Requested Block Size 4.0 MB
242
  Peak Block count 1
243
  Peak Allocated memory 2.2 MB
@@ -249,13 +246,13 @@ Memory Statistics:
249
  Used Block count 1
250
  Peak Allocated bytes 1.0 MB
251
  [ALLOC_TYPETREE_MAIN]
252
- Peak usage frame count: [0-1.0 KB]: 5466 frames
253
  Requested Block Size 2.0 MB
254
  Peak Block count 1
255
  Peak Allocated memory 1.0 KB
256
  Peak Large allocation bytes 0 B
257
  [ALLOC_TYPETREE_THREAD]
258
- Peak usage frame count: [1.0 KB-2.0 KB]: 5466 frames
259
  Requested Block Size 2.0 MB
260
  Peak Block count 1
261
  Peak Allocated memory 1.7 KB
 
2
  Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
5
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
  [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
 
31
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
  FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
+ - Completed reload, in 0.079 seconds
35
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
42
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
43
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
44
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
45
+ UnloadTime: 0.716327 ms
46
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
47
  requesting resize 84 x 84
48
  Setting up 1 worker threads for Enlighten.
 
50
  [ALLOC_TEMP_TLS] TLS Allocator
51
  StackAllocators :
52
  [ALLOC_TEMP_MAIN]
53
+ Peak usage frame count: [8.0 KB-16.0 KB]: 10741 frames, [16.0 KB-32.0 KB]: 182 frames, [2.0 MB-4.0 MB]: 1 frames
54
  Initial Block Size 4.0 MB
55
  Current Block Size 4.0 MB
56
  Peak Allocated Bytes 2.0 MB
 
166
  Peak Allocated Bytes 0 B
167
  Overflow Count 0
168
  [ALLOC_DEFAULT] Dual Thread Allocator
169
+ Peak main deferred allocation count 41
170
  [ALLOC_BUCKET]
171
  Large Block size 4.0 MB
172
  Used Block count 1
173
  Peak Allocated bytes 1.0 MB
174
  [ALLOC_DEFAULT_MAIN]
175
+ Peak usage frame count: [4.0 MB-8.0 MB]: 10918 frames, [8.0 MB-16.0 MB]: 6 frames
176
  Requested Block Size 16.0 MB
177
  Peak Block count 1
178
+ Peak Allocated memory 9.3 MB
179
  Peak Large allocation bytes 0 B
180
  [ALLOC_DEFAULT_THREAD]
181
+ Peak usage frame count: [16.0 MB-32.0 MB]: 10924 frames
182
  Requested Block Size 16.0 MB
183
  Peak Block count 1
184
+ Peak Allocated memory 17.3 MB
185
  Peak Large allocation bytes 16.0 MB
186
  [ALLOC_TEMP_JOB_1_FRAME]
187
  Initial Block Size 2.0 MB
 
210
  Used Block count 1
211
  Peak Allocated bytes 1.0 MB
212
  [ALLOC_GFX_MAIN]
213
+ Peak usage frame count: [32.0 KB-64.0 KB]: 10392 frames, [64.0 KB-128.0 KB]: 532 frames
214
  Requested Block Size 16.0 MB
215
  Peak Block count 1
216
+ Peak Allocated memory 66.4 KB
217
  Peak Large allocation bytes 0 B
218
  [ALLOC_GFX_THREAD]
219
+ Peak usage frame count: [32.0 KB-64.0 KB]: 10924 frames
220
  Requested Block Size 16.0 MB
221
  Peak Block count 1
222
  Peak Allocated memory 39.6 KB
 
228
  Used Block count 1
229
  Peak Allocated bytes 1.0 MB
230
  [ALLOC_CACHEOBJECTS_MAIN]
231
+ Peak usage frame count: [0.5 MB-1.0 MB]: 10924 frames
232
  Requested Block Size 4.0 MB
233
  Peak Block count 1
234
  Peak Allocated memory 0.6 MB
235
  Peak Large allocation bytes 0 B
236
  [ALLOC_CACHEOBJECTS_THREAD]
237
+ Peak usage frame count: [0.5 MB-1.0 MB]: 10923 frames, [2.0 MB-4.0 MB]: 1 frames
238
  Requested Block Size 4.0 MB
239
  Peak Block count 1
240
  Peak Allocated memory 2.2 MB
 
246
  Used Block count 1
247
  Peak Allocated bytes 1.0 MB
248
  [ALLOC_TYPETREE_MAIN]
249
+ Peak usage frame count: [0-1.0 KB]: 10924 frames
250
  Requested Block Size 2.0 MB
251
  Peak Block count 1
252
  Peak Allocated memory 1.0 KB
253
  Peak Large allocation bytes 0 B
254
  [ALLOC_TYPETREE_THREAD]
255
+ Peak usage frame count: [1.0 KB-2.0 KB]: 10924 frames
256
  Requested Block Size 2.0 MB
257
  Peak Block count 1
258
  Peak Allocated memory 1.7 KB
run_logs/timers.json CHANGED
@@ -2,39 +2,51 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.9074764847755432,
6
- "min": 0.8980207443237305,
7
- "max": 2.887725353240967,
8
  "count": 40
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 4102.701171875,
12
- "min": 4102.701171875,
13
- "max": 16009.548828125,
14
  "count": 40
15
  },
16
  "SnowballTarget.Step.mean": {
17
- "value": 199920.0,
18
- "min": 4976.0,
19
- "max": 199920.0,
20
  "count": 40
21
  },
22
  "SnowballTarget.Step.sum": {
23
- "value": 199920.0,
24
- "min": 4976.0,
25
- "max": 199920.0,
26
  "count": 40
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 12.52747631072998,
30
- "min": 0.09490196406841278,
31
- "max": 12.592220306396484,
32
  "count": 40
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 901.978271484375,
36
- "min": 6.83294153213501,
37
- "max": 1038.786865234375,
 
 
 
 
 
 
 
 
 
 
 
 
38
  "count": 40
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
@@ -44,93 +56,129 @@
44
  "count": 40
45
  },
46
  "SnowballTarget.Environment.EpisodeLength.sum": {
47
- "value": 6567.0,
48
- "min": 4378.0,
49
- "max": 6567.0,
50
  "count": 40
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.07068405189210727,
54
- "min": 0.06319188880186309,
55
- "max": 0.07711656137079818,
56
  "count": 40
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.14136810378421455,
60
- "min": 0.12638377760372618,
61
- "max": 0.22135205152216098,
62
  "count": 40
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.20135490622590568,
66
- "min": 0.10407538581091691,
67
- "max": 0.2827119831653202,
68
  "count": 40
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.40270981245181137,
72
- "min": 0.20815077162183382,
73
- "max": 0.7716041477287517,
74
  "count": 40
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
- "value": 4.782098406000002e-06,
78
- "min": 4.782098406000002e-06,
79
- "max": 0.000295182001606,
80
  "count": 40
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
- "value": 9.564196812000005e-06,
84
- "min": 9.564196812000005e-06,
85
- "max": 0.0008211960262679999,
86
  "count": 40
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
89
- "value": 0.10239100000000001,
90
- "min": 0.10239100000000001,
91
- "max": 0.24759100000000003,
92
  "count": 40
93
  },
94
  "SnowballTarget.Policy.Epsilon.sum": {
95
- "value": 0.20478200000000002,
96
- "min": 0.20478200000000002,
97
- "max": 0.7105980000000001,
98
  "count": 40
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
- "value": 8.954060000000004e-05,
102
- "min": 8.954060000000004e-05,
103
- "max": 0.004919860599999999,
104
  "count": 40
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
- "value": 0.00017908120000000008,
108
- "min": 0.00017908120000000008,
109
- "max": 0.0136892268,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
110
  "count": 40
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 24.772727272727273,
114
- "min": 2.8636363636363638,
115
- "max": 25.09090909090909,
116
  "count": 40
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 545.0,
120
- "min": 63.0,
121
- "max": 808.0,
122
  "count": 40
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 24.772727272727273,
126
- "min": 2.8636363636363638,
127
- "max": 25.09090909090909,
128
  "count": 40
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 545.0,
132
- "min": 63.0,
133
- "max": 808.0,
 
 
 
 
 
 
 
 
 
 
 
 
134
  "count": 40
135
  },
136
  "SnowballTarget.IsTraining.mean": {
@@ -148,67 +196,67 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1689914868",
152
  "python_version": "3.10.6 (main, May 29 2023, 11:10:38) [GCC 11.3.0]",
153
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./ppo_snowballtarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "0.31.0.dev0",
155
  "mlagents_envs_version": "0.31.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.11.0+cu102",
158
  "numpy_version": "1.21.2",
159
- "end_time_seconds": "1689915421"
160
  },
161
- "total": 552.590371831,
162
  "count": 1,
163
- "self": 0.8758308999999826,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.03300840200006405,
167
  "count": 1,
168
- "self": 0.03300840200006405
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 551.6815325289999,
172
  "count": 1,
173
- "self": 0.6336164739990409,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 6.078368554999997,
177
  "count": 1,
178
- "self": 6.078368554999997
179
  },
180
  "TrainerController.advance": {
181
- "total": 544.7398578980009,
182
- "count": 18218,
183
- "self": 0.30483696400312965,
184
  "children": {
185
  "env_step": {
186
- "total": 544.4350209339977,
187
- "count": 18218,
188
- "self": 397.6144213920046,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 146.50968025798602,
192
- "count": 18218,
193
- "self": 1.9840145679867192,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 144.5256656899993,
197
- "count": 18218,
198
- "self": 144.5256656899993
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.3109192840071273,
204
- "count": 18218,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 549.7212558960065,
209
- "count": 18218,
210
  "is_parallel": true,
211
- "self": 284.3714180130047,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +265,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.0063523060000534315,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.004795859000068958,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0015564469999844732,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0015564469999844732
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.0849465590000591,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0009313829999655354,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.0004525500000909233,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.0004525500000909233
244
  },
245
  "communicator.exchange": {
246
- "total": 0.08036101599998347,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.08036101599998347
250
  },
251
  "steps_from_proto": {
252
- "total": 0.0032016100000191727,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.0007062609996637548,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.002495349000355418,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.002495349000355418
262
  }
263
  }
264
  }
@@ -267,34 +315,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 265.3498378830018,
271
- "count": 18217,
272
  "is_parallel": true,
273
- "self": 11.18796804900171,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 5.470085376996963,
277
- "count": 18217,
278
  "is_parallel": true,
279
- "self": 5.470085376996963
280
  },
281
  "communicator.exchange": {
282
- "total": 212.13326891400357,
283
- "count": 18217,
284
  "is_parallel": true,
285
- "self": 212.13326891400357
286
  },
287
  "steps_from_proto": {
288
- "total": 36.558515542999544,
289
- "count": 18217,
290
  "is_parallel": true,
291
- "self": 7.096372885992309,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 29.462142657007234,
295
- "count": 182170,
296
  "is_parallel": true,
297
- "self": 29.462142657007234
298
  }
299
  }
300
  }
@@ -309,9 +357,9 @@
309
  }
310
  },
311
  "trainer_threads": {
312
- "total": 0.00027281200004836137,
313
  "count": 1,
314
- "self": 0.00027281200004836137,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
@@ -320,36 +368,36 @@
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
- "total": 541.8961593720343,
324
- "count": 354203,
325
  "is_parallel": true,
326
- "self": 7.870481475088809,
327
  "children": {
328
  "process_trajectory": {
329
- "total": 215.14901628794598,
330
- "count": 354203,
331
  "is_parallel": true,
332
- "self": 210.52944220694633,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
- "total": 4.619574080999655,
336
- "count": 20,
337
  "is_parallel": true,
338
- "self": 4.619574080999655
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
- "total": 318.8766616089995,
344
- "count": 90,
345
  "is_parallel": true,
346
- "self": 125.46813759000077,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
- "total": 193.40852401899872,
350
- "count": 7645,
351
  "is_parallel": true,
352
- "self": 193.40852401899872
353
  }
354
  }
355
  }
@@ -360,14 +408,14 @@
360
  }
361
  },
362
  "TrainerController._save_models": {
363
- "total": 0.2294167899999593,
364
  "count": 1,
365
- "self": 0.0010001840000768425,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
- "total": 0.22841660599988245,
369
  "count": 1,
370
- "self": 0.22841660599988245
371
  }
372
  }
373
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.7823862433433533,
6
+ "min": 0.7754074335098267,
7
+ "max": 2.87212872505188,
8
  "count": 40
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 6962.455078125,
12
+ "min": 6900.3505859375,
13
+ "max": 31340.669921875,
14
  "count": 40
15
  },
16
  "SnowballTarget.Step.mean": {
17
+ "value": 399928.0,
18
+ "min": 9952.0,
19
+ "max": 399928.0,
20
  "count": 40
21
  },
22
  "SnowballTarget.Step.sum": {
23
+ "value": 399928.0,
24
+ "min": 9952.0,
25
+ "max": 399928.0,
26
  "count": 40
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 12.034942626953125,
30
+ "min": 0.45372357964515686,
31
+ "max": 12.034942626953125,
32
  "count": 40
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 1733.03173828125,
36
+ "min": 65.33619689941406,
37
+ "max": 1863.9541015625,
38
+ "count": 40
39
+ },
40
+ "SnowballTarget.Policy.CuriosityValueEstimate.mean": {
41
+ "value": 0.705839991569519,
42
+ "min": 0.15340012311935425,
43
+ "max": 1.2888423204421997,
44
+ "count": 40
45
+ },
46
+ "SnowballTarget.Policy.CuriosityValueEstimate.sum": {
47
+ "value": 101.64096069335938,
48
+ "min": 22.089616775512695,
49
+ "max": 197.5901641845703,
50
  "count": 40
51
  },
52
  "SnowballTarget.Environment.EpisodeLength.mean": {
 
56
  "count": 40
57
  },
58
  "SnowballTarget.Environment.EpisodeLength.sum": {
59
+ "value": 10945.0,
60
+ "min": 8756.0,
61
+ "max": 10945.0,
62
  "count": 40
63
  },
64
  "SnowballTarget.Losses.PolicyLoss.mean": {
65
+ "value": 0.06742327443161604,
66
+ "min": 0.0635198512355975,
67
+ "max": 0.07569230777824379,
68
  "count": 40
69
  },
70
  "SnowballTarget.Losses.PolicyLoss.sum": {
71
+ "value": 0.26969309772646416,
72
+ "min": 0.25407940494239,
73
+ "max": 0.3784615388912189,
74
  "count": 40
75
  },
76
  "SnowballTarget.Losses.ValueLoss.mean": {
77
+ "value": 0.0851801988285254,
78
+ "min": 0.05186068983508877,
79
+ "max": 0.10859993385041461,
80
  "count": 40
81
  },
82
  "SnowballTarget.Losses.ValueLoss.sum": {
83
+ "value": 0.3407207953141016,
84
+ "min": 0.20744275934035508,
85
+ "max": 0.5272768591256702,
86
  "count": 40
87
  },
88
  "SnowballTarget.Policy.LearningRate.mean": {
89
+ "value": 3.825098724999996e-06,
90
+ "min": 3.825098724999996e-06,
91
+ "max": 0.000295941001353,
92
  "count": 40
93
  },
94
  "SnowballTarget.Policy.LearningRate.sum": {
95
+ "value": 1.5300394899999985e-05,
96
+ "min": 1.5300394899999985e-05,
97
+ "max": 0.00144225001925,
98
  "count": 40
99
  },
100
  "SnowballTarget.Policy.Epsilon.mean": {
101
+ "value": 0.10191249999999999,
102
+ "min": 0.10191249999999999,
103
+ "max": 0.2479705,
104
  "count": 40
105
  },
106
  "SnowballTarget.Policy.Epsilon.sum": {
107
+ "value": 0.40764999999999996,
108
+ "min": 0.40764999999999996,
109
+ "max": 1.221125,
110
  "count": 40
111
  },
112
  "SnowballTarget.Policy.Beta.mean": {
113
+ "value": 7.362249999999992e-05,
114
+ "min": 7.362249999999992e-05,
115
+ "max": 0.0049324853,
116
  "count": 40
117
  },
118
  "SnowballTarget.Policy.Beta.sum": {
119
+ "value": 0.0002944899999999997,
120
+ "min": 0.0002944899999999997,
121
+ "max": 0.024039425,
122
+ "count": 40
123
+ },
124
+ "SnowballTarget.Losses.CuriosityForwardLoss.mean": {
125
+ "value": 0.04188172720810946,
126
+ "min": 0.033935607860193534,
127
+ "max": 0.08911535578560742,
128
+ "count": 40
129
+ },
130
+ "SnowballTarget.Losses.CuriosityForwardLoss.sum": {
131
+ "value": 0.16752690883243784,
132
+ "min": 0.13837627082624857,
133
+ "max": 0.3564614231424297,
134
+ "count": 40
135
+ },
136
+ "SnowballTarget.Losses.CuriosityInverseLoss.mean": {
137
+ "value": 0.6003185132847113,
138
+ "min": 0.5939198171391207,
139
+ "max": 2.693057048539905,
140
+ "count": 40
141
+ },
142
+ "SnowballTarget.Losses.CuriosityInverseLoss.sum": {
143
+ "value": 2.4012740531388452,
144
+ "min": 2.4012740531388452,
145
+ "max": 11.969808614955229,
146
  "count": 40
147
  },
148
  "SnowballTarget.Environment.CumulativeReward.mean": {
149
+ "value": 25.931818181818183,
150
+ "min": 3.25,
151
+ "max": 26.204545454545453,
152
  "count": 40
153
  },
154
  "SnowballTarget.Environment.CumulativeReward.sum": {
155
+ "value": 1141.0,
156
+ "min": 143.0,
157
+ "max": 1441.0,
158
  "count": 40
159
  },
160
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
161
+ "value": 23.338635753501546,
162
+ "min": 2.9249998954209415,
163
+ "max": 23.584090276197955,
164
  "count": 40
165
  },
166
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
167
+ "value": 1026.899973154068,
168
+ "min": 128.69999539852142,
169
+ "max": 1296.8999617099762,
170
+ "count": 40
171
+ },
172
+ "SnowballTarget.Policy.CuriosityReward.mean": {
173
+ "value": 1.5325787596997211,
174
+ "min": 1.4327925314618783,
175
+ "max": 2.6246763768859886,
176
+ "count": 40
177
+ },
178
+ "SnowballTarget.Policy.CuriosityReward.sum": {
179
+ "value": 67.43346542678773,
180
+ "min": 63.73716417513788,
181
+ "max": 144.35720072872937,
182
  "count": 40
183
  },
184
  "SnowballTarget.IsTraining.mean": {
 
196
  },
197
  "metadata": {
198
  "timer_format_version": "0.1.0",
199
+ "start_time_seconds": "1689921905",
200
  "python_version": "3.10.6 (main, May 29 2023, 11:10:38) [GCC 11.3.0]",
201
+ "command_line_arguments": "/usr/local/bin/mlagents-learn ./ppo_snowballtarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTargetCurious --no-graphics --force",
202
  "mlagents_version": "0.31.0.dev0",
203
  "mlagents_envs_version": "0.31.0.dev0",
204
  "communication_protocol_version": "1.5.0",
205
  "pytorch_version": "1.11.0+cu102",
206
  "numpy_version": "1.21.2",
207
+ "end_time_seconds": "1689923075"
208
  },
209
+ "total": 1170.717425587999,
210
  "count": 1,
211
+ "self": 0.42497604499931185,
212
  "children": {
213
  "run_training.setup": {
214
+ "total": 0.03128317899972899,
215
  "count": 1,
216
+ "self": 0.03128317899972899
217
  },
218
  "TrainerController.start_learning": {
219
+ "total": 1170.261166364,
220
  "count": 1,
221
+ "self": 1.0667345250103608,
222
  "children": {
223
  "TrainerController._reset_env": {
224
+ "total": 4.5801261289998365,
225
  "count": 1,
226
+ "self": 4.5801261289998365
227
  },
228
  "TrainerController.advance": {
229
+ "total": 1164.4760683689901,
230
+ "count": 36412,
231
+ "self": 0.5050375901291773,
232
  "children": {
233
  "env_step": {
234
+ "total": 1163.971030778861,
235
+ "count": 36412,
236
+ "self": 873.7409455367024,
237
  "children": {
238
  "SubprocessEnvManager._take_step": {
239
+ "total": 289.68929279413896,
240
+ "count": 36412,
241
+ "self": 3.2346155361719866,
242
  "children": {
243
  "TorchPolicy.evaluate": {
244
+ "total": 286.454677257967,
245
+ "count": 36412,
246
+ "self": 286.454677257967
247
  }
248
  }
249
  },
250
  "workers": {
251
+ "total": 0.5407924480196016,
252
+ "count": 36412,
253
  "self": 0.0,
254
  "children": {
255
  "worker_root": {
256
+ "total": 1167.2430328119744,
257
+ "count": 36412,
258
  "is_parallel": true,
259
+ "self": 681.4198886719532,
260
  "children": {
261
  "run_training.setup": {
262
  "total": 0.0,
 
265
  "self": 0.0,
266
  "children": {
267
  "steps_from_proto": {
268
+ "total": 0.0018927660003100755,
269
  "count": 1,
270
  "is_parallel": true,
271
+ "self": 0.000562683999305591,
272
  "children": {
273
  "_process_rank_one_or_two_observation": {
274
+ "total": 0.0013300820010044845,
275
  "count": 10,
276
  "is_parallel": true,
277
+ "self": 0.0013300820010044845
278
  }
279
  }
280
  },
281
  "UnityEnvironment.step": {
282
+ "total": 0.06856268100000307,
283
  "count": 1,
284
  "is_parallel": true,
285
+ "self": 0.0005964209994999692,
286
  "children": {
287
  "UnityEnvironment._generate_step_input": {
288
+ "total": 0.0003943410001738812,
289
  "count": 1,
290
  "is_parallel": true,
291
+ "self": 0.0003943410001738812
292
  },
293
  "communicator.exchange": {
294
+ "total": 0.06546917000014219,
295
  "count": 1,
296
  "is_parallel": true,
297
+ "self": 0.06546917000014219
298
  },
299
  "steps_from_proto": {
300
+ "total": 0.002102749000187032,
301
  "count": 1,
302
  "is_parallel": true,
303
+ "self": 0.0005411560005086358,
304
  "children": {
305
  "_process_rank_one_or_two_observation": {
306
+ "total": 0.0015615929996783962,
307
  "count": 10,
308
  "is_parallel": true,
309
+ "self": 0.0015615929996783962
310
  }
311
  }
312
  }
 
315
  }
316
  },
317
  "UnityEnvironment.step": {
318
+ "total": 485.8231441400212,
319
+ "count": 36411,
320
  "is_parallel": true,
321
+ "self": 20.639413553023587,
322
  "children": {
323
  "UnityEnvironment._generate_step_input": {
324
+ "total": 10.078052166051748,
325
+ "count": 36411,
326
  "is_parallel": true,
327
+ "self": 10.078052166051748
328
  },
329
  "communicator.exchange": {
330
+ "total": 387.16746825984774,
331
+ "count": 36411,
332
  "is_parallel": true,
333
+ "self": 387.16746825984774
334
  },
335
  "steps_from_proto": {
336
+ "total": 67.93821016109814,
337
+ "count": 36411,
338
  "is_parallel": true,
339
+ "self": 12.569071423020432,
340
  "children": {
341
  "_process_rank_one_or_two_observation": {
342
+ "total": 55.36913873807771,
343
+ "count": 364110,
344
  "is_parallel": true,
345
+ "self": 55.36913873807771
346
  }
347
  }
348
  }
 
357
  }
358
  },
359
  "trainer_threads": {
360
+ "total": 0.00020994999977119733,
361
  "count": 1,
362
+ "self": 0.00020994999977119733,
363
  "children": {
364
  "thread_root": {
365
  "total": 0.0,
 
368
  "self": 0.0,
369
  "children": {
370
  "trainer_advance": {
371
+ "total": 1165.3188830539748,
372
+ "count": 22648,
373
  "is_parallel": true,
374
+ "self": 0.5316522920466014,
375
  "children": {
376
  "process_trajectory": {
377
+ "total": 165.31442269693162,
378
+ "count": 22648,
379
  "is_parallel": true,
380
+ "self": 155.09521240393315,
381
  "children": {
382
  "RLTrainer._checkpoint": {
383
+ "total": 10.219210292998469,
384
+ "count": 40,
385
  "is_parallel": true,
386
+ "self": 10.219210292998469
387
  }
388
  }
389
  },
390
  "_update_policy": {
391
+ "total": 999.4728080649966,
392
+ "count": 181,
393
  "is_parallel": true,
394
+ "self": 622.892439668005,
395
  "children": {
396
  "TorchPPOOptimizer.update": {
397
+ "total": 376.5803683969916,
398
+ "count": 15380,
399
  "is_parallel": true,
400
+ "self": 376.5803683969916
401
  }
402
  }
403
  }
 
408
  }
409
  },
410
  "TrainerController._save_models": {
411
+ "total": 0.13802739099992323,
412
  "count": 1,
413
+ "self": 0.0007361719999607885,
414
  "children": {
415
  "RLTrainer._checkpoint": {
416
+ "total": 0.13729121899996244,
417
  "count": 1,
418
+ "self": 0.13729121899996244
419
  }
420
  }
421
  }
run_logs/training_status.json CHANGED
@@ -2,103 +2,103 @@
2
  "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
- "steps": 119952,
6
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-119952.onnx",
7
- "reward": 21.454545454545453,
8
- "creation_time": 1689915197.0312555,
9
  "auxillary_file_paths": [
10
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-119952.pt"
11
  ]
12
  },
13
  {
14
- "steps": 129992,
15
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-129992.onnx",
16
- "reward": 22.636363636363637,
17
- "creation_time": 1689915226.046242,
18
  "auxillary_file_paths": [
19
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-129992.pt"
20
  ]
21
  },
22
  {
23
- "steps": 139944,
24
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-139944.onnx",
25
- "reward": 24.90909090909091,
26
- "creation_time": 1689915251.506197,
27
  "auxillary_file_paths": [
28
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-139944.pt"
29
  ]
30
  },
31
  {
32
- "steps": 149984,
33
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
34
- "reward": 23.818181818181817,
35
- "creation_time": 1689915279.6826274,
36
  "auxillary_file_paths": [
37
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
38
  ]
39
  },
40
  {
41
- "steps": 159936,
42
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-159936.onnx",
43
- "reward": 24.0,
44
- "creation_time": 1689915305.8110385,
45
  "auxillary_file_paths": [
46
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-159936.pt"
47
  ]
48
  },
49
  {
50
- "steps": 169976,
51
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-169976.onnx",
52
- "reward": 24.181818181818183,
53
- "creation_time": 1689915334.796858,
54
  "auxillary_file_paths": [
55
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-169976.pt"
56
  ]
57
  },
58
  {
59
- "steps": 179928,
60
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-179928.onnx",
61
- "reward": 25.454545454545453,
62
- "creation_time": 1689915363.8239555,
63
  "auxillary_file_paths": [
64
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-179928.pt"
65
  ]
66
  },
67
  {
68
- "steps": 189968,
69
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-189968.onnx",
70
- "reward": 24.0,
71
- "creation_time": 1689915393.1039462,
72
  "auxillary_file_paths": [
73
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-189968.pt"
74
  ]
75
  },
76
  {
77
- "steps": 199920,
78
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199920.onnx",
79
- "reward": 23.90909090909091,
80
- "creation_time": 1689915420.3405128,
81
  "auxillary_file_paths": [
82
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-199920.pt"
83
  ]
84
  },
85
  {
86
- "steps": 200112,
87
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
88
- "reward": 23.90909090909091,
89
- "creation_time": 1689915420.6127658,
90
  "auxillary_file_paths": [
91
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
92
  ]
93
  }
94
  ],
95
  "final_checkpoint": {
96
- "steps": 200112,
97
- "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
98
- "reward": 23.90909090909091,
99
- "creation_time": 1689915420.6127658,
100
  "auxillary_file_paths": [
101
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
102
  ]
103
  }
104
  },
 
2
  "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
+ "steps": 319960,
6
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-319960.onnx",
7
+ "reward": null,
8
+ "creation_time": 1689922845.003947,
9
  "auxillary_file_paths": [
10
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-319960.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 329992,
15
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-329992.onnx",
16
+ "reward": 26.3,
17
+ "creation_time": 1689922871.4831252,
18
  "auxillary_file_paths": [
19
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-329992.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 339952,
24
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-339952.onnx",
25
+ "reward": null,
26
+ "creation_time": 1689922902.8197052,
27
  "auxillary_file_paths": [
28
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-339952.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 349992,
33
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-349992.onnx",
34
+ "reward": null,
35
+ "creation_time": 1689922934.255856,
36
  "auxillary_file_paths": [
37
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-349992.pt"
38
  ]
39
  },
40
  {
41
+ "steps": 359944,
42
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-359944.onnx",
43
+ "reward": null,
44
+ "creation_time": 1689922960.0544336,
45
  "auxillary_file_paths": [
46
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-359944.pt"
47
  ]
48
  },
49
  {
50
+ "steps": 369984,
51
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-369984.onnx",
52
+ "reward": null,
53
+ "creation_time": 1689922991.6703765,
54
  "auxillary_file_paths": [
55
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-369984.pt"
56
  ]
57
  },
58
  {
59
+ "steps": 379936,
60
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-379936.onnx",
61
+ "reward": null,
62
+ "creation_time": 1689923017.6110141,
63
  "auxillary_file_paths": [
64
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-379936.pt"
65
  ]
66
  },
67
  {
68
+ "steps": 389976,
69
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-389976.onnx",
70
+ "reward": null,
71
+ "creation_time": 1689923048.6691453,
72
  "auxillary_file_paths": [
73
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-389976.pt"
74
  ]
75
  },
76
  {
77
+ "steps": 399928,
78
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-399928.onnx",
79
+ "reward": null,
80
+ "creation_time": 1689923075.3943675,
81
  "auxillary_file_paths": [
82
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-399928.pt"
83
  ]
84
  },
85
  {
86
+ "steps": 400120,
87
+ "file_path": "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-400120.onnx",
88
+ "reward": null,
89
+ "creation_time": 1689923075.5676386,
90
  "auxillary_file_paths": [
91
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-400120.pt"
92
  ]
93
  }
94
  ],
95
  "final_checkpoint": {
96
+ "steps": 400120,
97
+ "file_path": "results/SnowballTargetCurious/SnowballTarget.onnx",
98
+ "reward": null,
99
+ "creation_time": 1689923075.5676386,
100
  "auxillary_file_paths": [
101
+ "results/SnowballTargetCurious/SnowballTarget/SnowballTarget-400120.pt"
102
  ]
103
  }
104
  },