danieladejumo commited on
Commit
13c46c7
β€’
1 Parent(s): 9ca9d8f

Fixed bash bug - mlagents-learn not found

Browse files
Pyramids.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4348036dd0d5c6861bd2445b23f192edcfe446621e69c109af128946089888f
3
  size 1417437
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7354700265026d0b204f8d157916bd721a01f208d280e96ebb116070f2d5c44b
3
  size 1417437
Pyramids/{Pyramids-1252618.onnx β†’ Pyramids-1000082.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4348036dd0d5c6861bd2445b23f192edcfe446621e69c109af128946089888f
3
  size 1417437
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7354700265026d0b204f8d157916bd721a01f208d280e96ebb116070f2d5c44b
3
  size 1417437
Pyramids/{Pyramids-1252618.pt β†’ Pyramids-1000082.pt} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43de801de6483535789e6e275e2873fde5ba3230cfd9d5f40548b2b7a5f13d3a
3
- size 8652382
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c65d553cb663ec3a833b5ad87422b01923361b5738c9b28bddd0eb62c1e6fa16
3
+ size 8651427
Pyramids/{Pyramids-999994.onnx β†’ Pyramids-499875.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:514254baaac8e22299c2a57b577a99db4d45e2aaf04426998357d35bbbe9341d
3
  size 1417437
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c0cf07c4bc7b96415ae53a02e0d6aa5e50c223321b0e809c0a3d09af3a72e4d
3
  size 1417437
Pyramids/{Pyramids-499879.pt β†’ Pyramids-499875.pt} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0e808b57f8c1d12f54b216bc6f79416b88bb2bf300dab6c6cc26aedbd8e9d69
3
- size 8652382
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b87c77e1bf1f559175fdb710a1343b2fb50ab6dad32876aab80c6a0221b68f3b
3
+ size 8651427
Pyramids/{Pyramids-499879.onnx β†’ Pyramids-999954.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f524239e34ea9c75ce74b41d8944baa745d869772b2aaf6463d7478ee201857
3
  size 1417437
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7354700265026d0b204f8d157916bd721a01f208d280e96ebb116070f2d5c44b
3
  size 1417437
Pyramids/Pyramids-999954.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9147fd1358b46d71b7f7b4fcabd669e3019212b3ce4c0dabb5da1fb8322e15e6
3
+ size 8651427
Pyramids/Pyramids-999994.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:16fce4ebf2b9000204d6b00d8d9447cab722a2f59c15e67932d3d296322b1fab
3
- size 8652382
 
 
 
 
Pyramids/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43de801de6483535789e6e275e2873fde5ba3230cfd9d5f40548b2b7a5f13d3a
3
- size 8652382
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c65d553cb663ec3a833b5ad87422b01923361b5738c9b28bddd0eb62c1e6fa16
3
+ size 8651427
Pyramids/{events.out.tfevents.1658672915.410d80450fc1.212.0 β†’ events.out.tfevents.1661155235.0f67ab6d30c4.464.0} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bb8d676725e407ebfb51121059b22f93d9b30285643187d2d80a6f512a3e116
3
- size 397091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f040e61d546ed03fdfc11d17f5f37982965ada99d3400a95faf4cc814c03e72b
3
+ size 332946
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 64, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 5, "checkpoint_interval": 500000, "max_steps": 3000000, "time_horizon": 128, "summary_freq": 30000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./trained-envs-executables/linux/Pyramids/Pyramids", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Pyramids Training", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 64, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 5, "checkpoint_interval": 500000, "max_steps": 1000000, "time_horizon": 128, "summary_freq": 30000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./trained-envs-executables/linux/Pyramids/Pyramids", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Pyramids Training", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -49,7 +49,7 @@ behaviors:
49
  init_path: null
50
  keep_checkpoints: 5
51
  checkpoint_interval: 500000
52
- max_steps: 3000000
53
  time_horizon: 128
54
  summary_freq: 30000
55
  threaded: false
 
49
  init_path: null
50
  keep_checkpoints: 5
51
  checkpoint_interval: 500000
52
+ max_steps: 1000000
53
  time_horizon: 128
54
  summary_freq: 30000
55
  threaded: false
run_logs/Player-0.log CHANGED
@@ -1,7 +1,7 @@
1
  Mono path[0] = '/content/ml-agents/trained-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
  Mono config path = '/content/ml-agents/trained-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
  Found 1 interfaces on host : 0) 172.28.0.2
4
- Multi-casting "[IP] 172.28.0.2 [Port] 55494 [Flags] 2 [Guid] 3747311524 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.2) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
7
  PlayerPrefs - Creating folder: /root/.config/unity3d/Unity Technologies
@@ -36,7 +36,7 @@ ALSA lib pcm.c:2495:(snd_pcm_open_noupdate) Unknown PCM default
36
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
37
  FMOD initialized on nosound output
38
  Begin MonoManager ReloadAssembly
39
- - Completed reload, in 0.094 seconds
40
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
42
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -68,7 +68,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
68
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
69
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
70
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
71
- UnloadTime: 0.973858 ms
72
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
73
  requesting resize 84 x 84
74
  Setting up 1 worker threads for Enlighten.
@@ -76,7 +76,7 @@ PlayerConnection::CleanupMemory Statistics:
76
  [ALLOC_TEMP_TLS] TLS Allocator
77
  StackAllocators :
78
  [ALLOC_TEMP_MAIN]
79
- Peak usage frame count: [16.0 KB-32.0 KB]: 2189 frames, [32.0 KB-64.0 KB]: 16034 frames, [64.0 KB-128.0 KB]: 5280 frames, [2.0 MB-4.0 MB]: 1 frames
80
  Initial Block Size 4.0 MB
81
  Current Block Size 4.0 MB
82
  Peak Allocated Bytes 2.0 MB
@@ -126,17 +126,17 @@ PlayerConnection::CleanupMemory Statistics:
126
  Current Block Size 64.0 KB
127
  Peak Allocated Bytes 0 B
128
  Overflow Count 0
129
- [ALLOC_TEMP_Background Job.Worker 2]
130
  Initial Block Size 32.0 KB
131
  Current Block Size 32.0 KB
132
  Peak Allocated Bytes 0 B
133
  Overflow Count 0
134
- [ALLOC_TEMP_Background Job.Worker 15]
135
  Initial Block Size 32.0 KB
136
  Current Block Size 32.0 KB
137
  Peak Allocated Bytes 0 B
138
  Overflow Count 0
139
- [ALLOC_TEMP_Background Job.Worker 7]
140
  Initial Block Size 32.0 KB
141
  Current Block Size 32.0 KB
142
  Peak Allocated Bytes 0 B
@@ -149,7 +149,7 @@ PlayerConnection::CleanupMemory Statistics:
149
  [ALLOC_TEMP_Profiler.Dispatcher]
150
  Initial Block Size 64.0 KB
151
  Current Block Size 64.0 KB
152
- Peak Allocated Bytes 240 B
153
  Overflow Count 0
154
  [ALLOC_TEMP_Background Job.Worker 12]
155
  Initial Block Size 32.0 KB
@@ -207,28 +207,28 @@ PlayerConnection::CleanupMemory Statistics:
207
  Peak Allocated Bytes 128 B
208
  Overflow Count 0
209
  [ALLOC_MEMORYPROFILER]
210
- Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 23205 frames
211
  Requested Block Size 1.0 MB
212
  Peak Block count 2
213
  Peak Allocated memory 1.6 MB
214
  Peak Large allocation bytes 0 B
215
  [ALLOC_DEFAULT] Dual Thread Allocator
216
- Peak main deferred allocation count 12856
217
  [ALLOC_BUCKET]
218
  Large Block size 4.0 MB
219
  Used Block count 1
220
- Peak Allocated bytes 1.5 MB
221
  [ALLOC_DEFAULT_MAIN]
222
- Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 23503 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 3
225
- Peak Allocated memory 31.0 MB
226
  Peak Large allocation bytes 0 B
227
  [ALLOC_DEFAULT_THREAD]
228
- Peak usage frame count: [16.0 MB-32.0 MB]: 23504 frames
229
  Requested Block Size 16.0 MB
230
  Peak Block count 1
231
- Peak Allocated memory 22.9 MB
232
  Peak Large allocation bytes 16.0 MB
233
  [ALLOC_TEMP_JOB_1_FRAME]
234
  Initial Block Size 2.0 MB
@@ -255,15 +255,15 @@ PlayerConnection::CleanupMemory Statistics:
255
  [ALLOC_BUCKET]
256
  Large Block size 4.0 MB
257
  Used Block count 1
258
- Peak Allocated bytes 1.5 MB
259
  [ALLOC_GFX_MAIN]
260
- Peak usage frame count: [32.0 KB-64.0 KB]: 23503 frames, [64.0 KB-128.0 KB]: 1 frames
261
  Requested Block Size 16.0 MB
262
  Peak Block count 1
263
  Peak Allocated memory 67.0 KB
264
  Peak Large allocation bytes 0 B
265
  [ALLOC_GFX_THREAD]
266
- Peak usage frame count: [64.0 KB-128.0 KB]: 23504 frames
267
  Requested Block Size 16.0 MB
268
  Peak Block count 1
269
  Peak Allocated memory 67.7 KB
@@ -273,15 +273,15 @@ PlayerConnection::CleanupMemory Statistics:
273
  [ALLOC_BUCKET]
274
  Large Block size 4.0 MB
275
  Used Block count 1
276
- Peak Allocated bytes 1.5 MB
277
  [ALLOC_CACHEOBJECTS_MAIN]
278
- Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 23205 frames
279
  Requested Block Size 4.0 MB
280
  Peak Block count 1
281
  Peak Allocated memory 1.4 MB
282
  Peak Large allocation bytes 0 B
283
  [ALLOC_CACHEOBJECTS_THREAD]
284
- Peak usage frame count: [1.0 MB-2.0 MB]: 23503 frames, [4.0 MB-8.0 MB]: 1 frames
285
  Requested Block Size 4.0 MB
286
  Peak Block count 2
287
  Peak Allocated memory 4.7 MB
@@ -291,27 +291,27 @@ PlayerConnection::CleanupMemory Statistics:
291
  [ALLOC_BUCKET]
292
  Large Block size 4.0 MB
293
  Used Block count 1
294
- Peak Allocated bytes 1.5 MB
295
  [ALLOC_TYPETREE_MAIN]
296
- Peak usage frame count: [0-1.0 KB]: 23504 frames
297
  Requested Block Size 2.0 MB
298
  Peak Block count 1
299
  Peak Allocated memory 1.0 KB
300
  Peak Large allocation bytes 0 B
301
  [ALLOC_TYPETREE_THREAD]
302
- Peak usage frame count: [1.0 KB-2.0 KB]: 23504 frames
303
  Requested Block Size 2.0 MB
304
  Peak Block count 1
305
  Peak Allocated memory 1.9 KB
306
  Peak Large allocation bytes 0 B
307
  [ALLOC_PROFILER]
308
- Peak usage frame count: [16.0 KB-32.0 KB]: 23504 frames
309
  Requested Block Size 16.0 MB
310
  Peak Block count 1
311
- Peak Allocated memory 29.2 KB
312
  Peak Large allocation bytes 0 B
313
  [ALLOC_PROFILER_BUCKET]
314
  Large Block size 4.0 MB
315
  Used Block count 1
316
  Peak Allocated bytes 396 B
317
- ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1658675698793,"processId":224,"allocatedMemory":1957329,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3444},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
 
1
  Mono path[0] = '/content/ml-agents/trained-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
  Mono config path = '/content/ml-agents/trained-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
  Found 1 interfaces on host : 0) 172.28.0.2
4
+ Multi-casting "[IP] 172.28.0.2 [Port] 55500 [Flags] 2 [Guid] 4225727653 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.2) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
7
  PlayerPrefs - Creating folder: /root/.config/unity3d/Unity Technologies
 
36
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
37
  FMOD initialized on nosound output
38
  Begin MonoManager ReloadAssembly
39
+ - Completed reload, in 0.114 seconds
40
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
42
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
68
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
69
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
70
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
71
+ UnloadTime: 1.321830 ms
72
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
73
  requesting resize 84 x 84
74
  Setting up 1 worker threads for Enlighten.
 
76
  [ALLOC_TEMP_TLS] TLS Allocator
77
  StackAllocators :
78
  [ALLOC_TEMP_MAIN]
79
+ Peak usage frame count: [16.0 KB-32.0 KB]: 4805 frames, [32.0 KB-64.0 KB]: 11876 frames, [64.0 KB-128.0 KB]: 2089 frames, [2.0 MB-4.0 MB]: 1 frames
80
  Initial Block Size 4.0 MB
81
  Current Block Size 4.0 MB
82
  Peak Allocated Bytes 2.0 MB
 
126
  Current Block Size 64.0 KB
127
  Peak Allocated Bytes 0 B
128
  Overflow Count 0
129
+ [ALLOC_TEMP_Background Job.Worker 15]
130
  Initial Block Size 32.0 KB
131
  Current Block Size 32.0 KB
132
  Peak Allocated Bytes 0 B
133
  Overflow Count 0
134
+ [ALLOC_TEMP_Background Job.Worker 7]
135
  Initial Block Size 32.0 KB
136
  Current Block Size 32.0 KB
137
  Peak Allocated Bytes 0 B
138
  Overflow Count 0
139
+ [ALLOC_TEMP_Background Job.Worker 2]
140
  Initial Block Size 32.0 KB
141
  Current Block Size 32.0 KB
142
  Peak Allocated Bytes 0 B
 
149
  [ALLOC_TEMP_Profiler.Dispatcher]
150
  Initial Block Size 64.0 KB
151
  Current Block Size 64.0 KB
152
+ Peak Allocated Bytes 480 B
153
  Overflow Count 0
154
  [ALLOC_TEMP_Background Job.Worker 12]
155
  Initial Block Size 32.0 KB
 
207
  Peak Allocated Bytes 128 B
208
  Overflow Count 0
209
  [ALLOC_MEMORYPROFILER]
210
+ Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 18472 frames
211
  Requested Block Size 1.0 MB
212
  Peak Block count 2
213
  Peak Allocated memory 1.6 MB
214
  Peak Large allocation bytes 0 B
215
  [ALLOC_DEFAULT] Dual Thread Allocator
216
+ Peak main deferred allocation count 12983
217
  [ALLOC_BUCKET]
218
  Large Block size 4.0 MB
219
  Used Block count 1
220
+ Peak Allocated bytes 1.6 MB
221
  [ALLOC_DEFAULT_MAIN]
222
+ Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 18770 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 3
225
+ Peak Allocated memory 30.8 MB
226
  Peak Large allocation bytes 0 B
227
  [ALLOC_DEFAULT_THREAD]
228
+ Peak usage frame count: [16.0 MB-32.0 MB]: 18771 frames
229
  Requested Block Size 16.0 MB
230
  Peak Block count 1
231
+ Peak Allocated memory 22.8 MB
232
  Peak Large allocation bytes 16.0 MB
233
  [ALLOC_TEMP_JOB_1_FRAME]
234
  Initial Block Size 2.0 MB
 
255
  [ALLOC_BUCKET]
256
  Large Block size 4.0 MB
257
  Used Block count 1
258
+ Peak Allocated bytes 1.6 MB
259
  [ALLOC_GFX_MAIN]
260
+ Peak usage frame count: [32.0 KB-64.0 KB]: 18770 frames, [64.0 KB-128.0 KB]: 1 frames
261
  Requested Block Size 16.0 MB
262
  Peak Block count 1
263
  Peak Allocated memory 67.0 KB
264
  Peak Large allocation bytes 0 B
265
  [ALLOC_GFX_THREAD]
266
+ Peak usage frame count: [64.0 KB-128.0 KB]: 18771 frames
267
  Requested Block Size 16.0 MB
268
  Peak Block count 1
269
  Peak Allocated memory 67.7 KB
 
273
  [ALLOC_BUCKET]
274
  Large Block size 4.0 MB
275
  Used Block count 1
276
+ Peak Allocated bytes 1.6 MB
277
  [ALLOC_CACHEOBJECTS_MAIN]
278
+ Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 18472 frames
279
  Requested Block Size 4.0 MB
280
  Peak Block count 1
281
  Peak Allocated memory 1.4 MB
282
  Peak Large allocation bytes 0 B
283
  [ALLOC_CACHEOBJECTS_THREAD]
284
+ Peak usage frame count: [1.0 MB-2.0 MB]: 18770 frames, [4.0 MB-8.0 MB]: 1 frames
285
  Requested Block Size 4.0 MB
286
  Peak Block count 2
287
  Peak Allocated memory 4.7 MB
 
291
  [ALLOC_BUCKET]
292
  Large Block size 4.0 MB
293
  Used Block count 1
294
+ Peak Allocated bytes 1.6 MB
295
  [ALLOC_TYPETREE_MAIN]
296
+ Peak usage frame count: [0-1.0 KB]: 18771 frames
297
  Requested Block Size 2.0 MB
298
  Peak Block count 1
299
  Peak Allocated memory 1.0 KB
300
  Peak Large allocation bytes 0 B
301
  [ALLOC_TYPETREE_THREAD]
302
+ Peak usage frame count: [1.0 KB-2.0 KB]: 18771 frames
303
  Requested Block Size 2.0 MB
304
  Peak Block count 1
305
  Peak Allocated memory 1.9 KB
306
  Peak Large allocation bytes 0 B
307
  [ALLOC_PROFILER]
308
+ Peak usage frame count: [16.0 KB-32.0 KB]: 18771 frames
309
  Requested Block Size 16.0 MB
310
  Peak Block count 1
311
+ Peak Allocated memory 29.0 KB
312
  Peak Large allocation bytes 0 B
313
  [ALLOC_PROFILER_BUCKET]
314
  Large Block size 4.0 MB
315
  Used Block count 1
316
  Peak Allocated bytes 396 B
317
+ ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1661158307785,"processId":474,"allocatedMemory":1957329,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3444},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
run_logs/timers.json CHANGED
@@ -2,189 +2,189 @@
2
  "name": "root",
3
  "gauges": {
4
  "Pyramids.Policy.Entropy.mean": {
5
- "value": 0.2899654805660248,
6
- "min": 0.28584811091423035,
7
- "max": 1.4236252307891846,
8
- "count": 41
9
  },
10
  "Pyramids.Policy.Entropy.sum": {
11
- "value": 8787.1142578125,
12
- "min": 8520.560546875,
13
- "max": 43187.09375,
14
- "count": 41
15
  },
16
  "Pyramids.Step.mean": {
17
- "value": 1229997.0,
18
  "min": 29952.0,
19
- "max": 1229997.0,
20
- "count": 41
21
  },
22
  "Pyramids.Step.sum": {
23
- "value": 1229997.0,
24
  "min": 29952.0,
25
- "max": 1229997.0,
26
- "count": 41
27
  },
28
  "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 0.8069287538528442,
30
- "min": -0.10137598216533661,
31
- "max": 0.8069287538528442,
32
- "count": 41
33
  },
34
  "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 240.4647674560547,
36
- "min": -24.330236434936523,
37
- "max": 240.4647674560547,
38
- "count": 41
39
  },
40
  "Pyramids.Policy.RndValueEstimate.mean": {
41
- "value": -0.013106937520205975,
42
- "min": -0.01805024966597557,
43
- "max": 0.4306671619415283,
44
- "count": 41
45
  },
46
  "Pyramids.Policy.RndValueEstimate.sum": {
47
- "value": -3.905867338180542,
48
- "min": -4.999919414520264,
49
- "max": 102.068115234375,
50
- "count": 41
51
  },
52
  "Pyramids.Losses.PolicyLoss.mean": {
53
- "value": 0.06829092282174463,
54
- "min": 0.06448111323734644,
55
- "max": 0.07420187728008157,
56
- "count": 41
57
  },
58
  "Pyramids.Losses.PolicyLoss.sum": {
59
- "value": 0.9560729195044249,
60
- "min": 0.4824396216002127,
61
- "max": 1.0618449050622683,
62
- "count": 41
63
  },
64
  "Pyramids.Losses.ValueLoss.mean": {
65
- "value": 0.012827238258300804,
66
- "min": 0.00045624834884924574,
67
- "max": 0.01638201445945035,
68
- "count": 41
69
  },
70
  "Pyramids.Losses.ValueLoss.sum": {
71
- "value": 0.17958133561621126,
72
- "min": 0.004106235139643212,
73
- "max": 0.2293482024323049,
74
- "count": 41
75
  },
76
  "Pyramids.Policy.LearningRate.mean": {
77
- "value": 0.00017846274051243338,
78
- "min": 0.00017846274051243338,
79
- "max": 0.00029838354339596195,
80
- "count": 41
81
  },
82
  "Pyramids.Policy.LearningRate.sum": {
83
- "value": 0.0024984783671740673,
84
- "min": 0.0020886848037717336,
85
- "max": 0.0038926744024419336,
86
- "count": 41
87
  },
88
  "Pyramids.Policy.Epsilon.mean": {
89
- "value": 0.1594875666666667,
90
- "min": 0.1594875666666667,
91
- "max": 0.19946118095238097,
92
- "count": 41
93
  },
94
  "Pyramids.Policy.Epsilon.sum": {
95
- "value": 2.232825933333334,
96
- "min": 1.3962282666666668,
97
- "max": 2.797558066666667,
98
- "count": 41
99
  },
100
  "Pyramids.Policy.Beta.mean": {
101
- "value": 0.005952807910000001,
102
- "min": 0.005952807910000001,
103
- "max": 0.009946171977142856,
104
- "count": 41
105
  },
106
  "Pyramids.Policy.Beta.sum": {
107
- "value": 0.08333931074,
108
- "min": 0.06962320384,
109
- "max": 0.12977605086,
110
- "count": 41
111
  },
112
  "Pyramids.Losses.RNDLoss.mean": {
113
- "value": 0.008461430668830872,
114
- "min": 0.008461430668830872,
115
- "max": 0.41180092096328735,
116
- "count": 41
117
  },
118
  "Pyramids.Losses.RNDLoss.sum": {
119
- "value": 0.1184600293636322,
120
- "min": 0.1184600293636322,
121
- "max": 2.8826065063476562,
122
- "count": 41
123
  },
124
  "Pyramids.Environment.EpisodeLength.mean": {
125
- "value": 237.584,
126
- "min": 237.584,
127
  "max": 999.0,
128
- "count": 41
129
  },
130
  "Pyramids.Environment.EpisodeLength.sum": {
131
- "value": 29698.0,
132
  "min": 15984.0,
133
- "max": 32502.0,
134
- "count": 41
135
  },
136
  "Pyramids.Environment.CumulativeReward.mean": {
137
- "value": 1.746411183834076,
138
  "min": -1.0000000521540642,
139
- "max": 1.746411183834076,
140
- "count": 41
141
  },
142
  "Pyramids.Environment.CumulativeReward.sum": {
143
- "value": 218.3013979792595,
144
- "min": -31.99920167028904,
145
- "max": 218.3013979792595,
146
- "count": 41
147
  },
148
  "Pyramids.Policy.ExtrinsicReward.mean": {
149
- "value": 1.746411183834076,
150
  "min": -1.0000000521540642,
151
- "max": 1.746411183834076,
152
- "count": 41
153
  },
154
  "Pyramids.Policy.ExtrinsicReward.sum": {
155
- "value": 218.3013979792595,
156
- "min": -31.99920167028904,
157
- "max": 218.3013979792595,
158
- "count": 41
159
  },
160
  "Pyramids.Policy.RndReward.mean": {
161
- "value": 0.020895628459169528,
162
- "min": 0.020895628459169528,
163
- "max": 8.422627033665776,
164
- "count": 41
165
  },
166
  "Pyramids.Policy.RndReward.sum": {
167
- "value": 2.611953557396191,
168
- "min": 2.587342332233675,
169
- "max": 134.76203253865242,
170
- "count": 41
171
  },
172
  "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
- "count": 41
177
  },
178
  "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
- "count": 41
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
- "start_time_seconds": "1658672908",
188
  "python_version": "3.7.13 (default, Apr 24 2022, 01:04:09) \n[GCC 7.5.0]",
189
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./trained-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
190
  "mlagents_version": "0.29.0.dev0",
@@ -192,66 +192,66 @@
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "1.8.1+cu102",
194
  "numpy_version": "1.21.6",
195
- "end_time_seconds": "1658675698"
196
  },
197
- "total": 2790.626398085,
198
  "count": 1,
199
- "self": 0.3481506959997205,
200
  "children": {
201
  "run_training.setup": {
202
- "total": 0.044852698999989116,
203
  "count": 1,
204
- "self": 0.044852698999989116
205
  },
206
  "TrainerController.start_learning": {
207
- "total": 2790.2333946900003,
208
  "count": 1,
209
- "self": 1.8101129170158856,
210
  "children": {
211
  "TrainerController._reset_env": {
212
- "total": 10.64714840000002,
213
  "count": 1,
214
- "self": 10.64714840000002
215
  },
216
  "TrainerController.advance": {
217
- "total": 2777.6252222529843,
218
- "count": 80700,
219
- "self": 1.9055193179219714,
220
  "children": {
221
  "env_step": {
222
- "total": 1844.4852667560303,
223
- "count": 80700,
224
- "self": 1707.0533900391108,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
- "total": 136.47011882598073,
228
- "count": 80700,
229
- "self": 5.937653609954424,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
- "total": 130.5324652160263,
233
- "count": 78346,
234
- "self": 44.758218357022855,
235
  "children": {
236
  "TorchPolicy.sample_actions": {
237
- "total": 85.77424685900345,
238
- "count": 78346,
239
- "self": 85.77424685900345
240
  }
241
  }
242
  }
243
  }
244
  },
245
  "workers": {
246
- "total": 0.9617578909388556,
247
- "count": 80700,
248
  "self": 0.0,
249
  "children": {
250
  "worker_root": {
251
- "total": 2785.100852380008,
252
- "count": 80700,
253
  "is_parallel": true,
254
- "self": 1209.7662778620079,
255
  "children": {
256
  "run_training.setup": {
257
  "total": 0.0,
@@ -260,48 +260,48 @@
260
  "self": 0.0,
261
  "children": {
262
  "steps_from_proto": {
263
- "total": 0.005518131000002313,
264
  "count": 1,
265
  "is_parallel": true,
266
- "self": 0.004134050999994088,
267
  "children": {
268
  "_process_rank_one_or_two_observation": {
269
- "total": 0.0013840800000082254,
270
  "count": 8,
271
  "is_parallel": true,
272
- "self": 0.0013840800000082254
273
  }
274
  }
275
  },
276
  "UnityEnvironment.step": {
277
- "total": 0.06052774400001226,
278
  "count": 1,
279
  "is_parallel": true,
280
- "self": 0.0005405889999963165,
281
  "children": {
282
  "UnityEnvironment._generate_step_input": {
283
- "total": 0.00047689000001582826,
284
  "count": 1,
285
  "is_parallel": true,
286
- "self": 0.00047689000001582826
287
  },
288
  "communicator.exchange": {
289
- "total": 0.056748214999998936,
290
  "count": 1,
291
  "is_parallel": true,
292
- "self": 0.056748214999998936
293
  },
294
  "steps_from_proto": {
295
- "total": 0.0027620500000011816,
296
  "count": 1,
297
  "is_parallel": true,
298
- "self": 0.00044432400000005146,
299
  "children": {
300
  "_process_rank_one_or_two_observation": {
301
- "total": 0.00231772600000113,
302
  "count": 8,
303
  "is_parallel": true,
304
- "self": 0.00231772600000113
305
  }
306
  }
307
  }
@@ -310,34 +310,34 @@
310
  }
311
  },
312
  "UnityEnvironment.step": {
313
- "total": 1575.334574518,
314
- "count": 80699,
315
  "is_parallel": true,
316
- "self": 36.3935605940203,
317
  "children": {
318
  "UnityEnvironment._generate_step_input": {
319
- "total": 31.00779965998197,
320
- "count": 80699,
321
  "is_parallel": true,
322
- "self": 31.00779965998197
323
  },
324
  "communicator.exchange": {
325
- "total": 1385.5735435379884,
326
- "count": 80699,
327
  "is_parallel": true,
328
- "self": 1385.5735435379884
329
  },
330
  "steps_from_proto": {
331
- "total": 122.35967072600923,
332
- "count": 80699,
333
  "is_parallel": true,
334
- "self": 30.59839916308505,
335
  "children": {
336
  "_process_rank_one_or_two_observation": {
337
- "total": 91.76127156292418,
338
- "count": 645592,
339
  "is_parallel": true,
340
- "self": 91.76127156292418
341
  }
342
  }
343
  }
@@ -350,31 +350,31 @@
350
  }
351
  },
352
  "trainer_advance": {
353
- "total": 931.234436179032,
354
- "count": 80700,
355
- "self": 3.438551608120406,
356
  "children": {
357
  "process_trajectory": {
358
- "total": 211.33477325291432,
359
- "count": 80700,
360
- "self": 211.12194592391432,
361
  "children": {
362
  "RLTrainer._checkpoint": {
363
- "total": 0.21282732899999246,
364
  "count": 2,
365
- "self": 0.21282732899999246
366
  }
367
  }
368
  },
369
  "_update_policy": {
370
- "total": 716.4611113179972,
371
- "count": 566,
372
- "self": 284.8564335240344,
373
  "children": {
374
  "TorchPPOOptimizer.update": {
375
- "total": 431.60467779396276,
376
- "count": 28588,
377
- "self": 431.60467779396276
378
  }
379
  }
380
  }
@@ -383,19 +383,19 @@
383
  }
384
  },
385
  "trainer_threads": {
386
- "total": 1.555000380903948e-06,
387
  "count": 1,
388
- "self": 1.555000380903948e-06
389
  },
390
  "TrainerController._save_models": {
391
- "total": 0.1509095649998926,
392
  "count": 1,
393
- "self": 0.002334846999929141,
394
  "children": {
395
  "RLTrainer._checkpoint": {
396
- "total": 0.14857471799996347,
397
  "count": 1,
398
- "self": 0.14857471799996347
399
  }
400
  }
401
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "Pyramids.Policy.Entropy.mean": {
5
+ "value": 0.537936270236969,
6
+ "min": 0.537936270236969,
7
+ "max": 1.3857479095458984,
8
+ "count": 33
9
  },
10
  "Pyramids.Policy.Entropy.sum": {
11
+ "value": 16301.6201171875,
12
+ "min": 16301.6201171875,
13
+ "max": 42038.046875,
14
+ "count": 33
15
  },
16
  "Pyramids.Step.mean": {
17
+ "value": 989974.0,
18
  "min": 29952.0,
19
+ "max": 989974.0,
20
+ "count": 33
21
  },
22
  "Pyramids.Step.sum": {
23
+ "value": 989974.0,
24
  "min": 29952.0,
25
+ "max": 989974.0,
26
+ "count": 33
27
  },
28
  "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 0.2987557351589203,
30
+ "min": -0.10508545488119125,
31
+ "max": 0.3129414916038513,
32
+ "count": 33
33
  },
34
  "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 77.37773895263672,
36
+ "min": -25.430679321289062,
37
+ "max": 80.42596435546875,
38
+ "count": 33
39
  },
40
  "Pyramids.Policy.RndValueEstimate.mean": {
41
+ "value": -0.016964631155133247,
42
+ "min": -0.016964631155133247,
43
+ "max": 0.3806672692298889,
44
+ "count": 33
45
  },
46
  "Pyramids.Policy.RndValueEstimate.sum": {
47
+ "value": -4.393839359283447,
48
+ "min": -4.393839359283447,
49
+ "max": 90.2181396484375,
50
+ "count": 33
51
  },
52
  "Pyramids.Losses.PolicyLoss.mean": {
53
+ "value": 0.06988025348853423,
54
+ "min": 0.06411768943503847,
55
+ "max": 0.07604630539601967,
56
+ "count": 33
57
  },
58
  "Pyramids.Losses.PolicyLoss.sum": {
59
+ "value": 1.0482038023280134,
60
+ "min": 0.5323241377721377,
61
+ "max": 1.0996024134733222,
62
+ "count": 33
63
  },
64
  "Pyramids.Losses.ValueLoss.mean": {
65
+ "value": 0.012970652159452178,
66
+ "min": 0.0006412699549866713,
67
+ "max": 0.014323089758630992,
68
+ "count": 33
69
  },
70
  "Pyramids.Losses.ValueLoss.sum": {
71
+ "value": 0.19455978239178268,
72
+ "min": 0.008977779369813399,
73
+ "max": 0.2005232566208339,
74
+ "count": 33
75
  },
76
  "Pyramids.Policy.LearningRate.mean": {
77
+ "value": 7.422677525806665e-06,
78
+ "min": 7.422677525806665e-06,
79
+ "max": 0.00029515063018788575,
80
+ "count": 33
81
  },
82
  "Pyramids.Policy.LearningRate.sum": {
83
+ "value": 0.00011134016288709997,
84
+ "min": 0.00011134016288709997,
85
+ "max": 0.0035084615305128993,
86
+ "count": 33
87
  },
88
  "Pyramids.Policy.Epsilon.mean": {
89
+ "value": 0.10247419333333334,
90
+ "min": 0.10247419333333334,
91
+ "max": 0.19838354285714285,
92
+ "count": 33
93
  },
94
  "Pyramids.Policy.Epsilon.sum": {
95
+ "value": 1.5371129000000001,
96
+ "min": 1.3886848,
97
+ "max": 2.5694871,
98
+ "count": 33
99
  },
100
  "Pyramids.Policy.Beta.mean": {
101
+ "value": 0.000257171914,
102
+ "min": 0.000257171914,
103
+ "max": 0.00983851593142857,
104
+ "count": 33
105
  },
106
  "Pyramids.Policy.Beta.sum": {
107
+ "value": 0.0038575787099999997,
108
+ "min": 0.0038575787099999997,
109
+ "max": 0.11697176129,
110
+ "count": 33
111
  },
112
  "Pyramids.Losses.RNDLoss.mean": {
113
+ "value": 0.012460886500775814,
114
+ "min": 0.012460886500775814,
115
+ "max": 0.5152274966239929,
116
+ "count": 33
117
  },
118
  "Pyramids.Losses.RNDLoss.sum": {
119
+ "value": 0.18691329658031464,
120
+ "min": 0.17502087354660034,
121
+ "max": 3.6065926551818848,
122
+ "count": 33
123
  },
124
  "Pyramids.Environment.EpisodeLength.mean": {
125
+ "value": 496.08620689655174,
126
+ "min": 496.08620689655174,
127
  "max": 999.0,
128
+ "count": 33
129
  },
130
  "Pyramids.Environment.EpisodeLength.sum": {
131
+ "value": 28773.0,
132
  "min": 15984.0,
133
+ "max": 32439.0,
134
+ "count": 33
135
  },
136
  "Pyramids.Environment.CumulativeReward.mean": {
137
+ "value": 1.3372526048568256,
138
  "min": -1.0000000521540642,
139
+ "max": 1.3372526048568256,
140
+ "count": 33
141
  },
142
  "Pyramids.Environment.CumulativeReward.sum": {
143
+ "value": 76.22339847683907,
144
+ "min": -30.469401702284813,
145
+ "max": 76.22339847683907,
146
+ "count": 33
147
  },
148
  "Pyramids.Policy.ExtrinsicReward.mean": {
149
+ "value": 1.3372526048568256,
150
  "min": -1.0000000521540642,
151
+ "max": 1.3372526048568256,
152
+ "count": 33
153
  },
154
  "Pyramids.Policy.ExtrinsicReward.sum": {
155
+ "value": 76.22339847683907,
156
+ "min": -30.469401702284813,
157
+ "max": 76.22339847683907,
158
+ "count": 33
159
  },
160
  "Pyramids.Policy.RndReward.mean": {
161
+ "value": 0.06230818303741041,
162
+ "min": 0.06230818303741041,
163
+ "max": 10.663450403138995,
164
+ "count": 33
165
  },
166
  "Pyramids.Policy.RndReward.sum": {
167
+ "value": 3.5515664331323933,
168
+ "min": 3.5515664331323933,
169
+ "max": 170.61520645022392,
170
+ "count": 33
171
  },
172
  "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
+ "count": 33
177
  },
178
  "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
+ "count": 33
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
+ "start_time_seconds": "1661155224",
188
  "python_version": "3.7.13 (default, Apr 24 2022, 01:04:09) \n[GCC 7.5.0]",
189
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./trained-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
190
  "mlagents_version": "0.29.0.dev0",
 
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "1.8.1+cu102",
194
  "numpy_version": "1.21.6",
195
+ "end_time_seconds": "1661158307"
196
  },
197
+ "total": 3083.676767436,
198
  "count": 1,
199
+ "self": 0.6833655189998353,
200
  "children": {
201
  "run_training.setup": {
202
+ "total": 0.05427349999990838,
203
  "count": 1,
204
+ "self": 0.05427349999990838
205
  },
206
  "TrainerController.start_learning": {
207
+ "total": 3082.939128417,
208
  "count": 1,
209
+ "self": 2.4252582980629995,
210
  "children": {
211
  "TrainerController._reset_env": {
212
+ "total": 11.044875691000016,
213
  "count": 1,
214
+ "self": 11.044875691000016
215
  },
216
  "TrainerController.advance": {
217
+ "total": 3069.361343868938,
218
+ "count": 63433,
219
+ "self": 2.6793028398874412,
220
  "children": {
221
  "env_step": {
222
+ "total": 1802.4806044610023,
223
+ "count": 63433,
224
+ "self": 1658.497047256109,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
+ "total": 142.53790752391296,
228
+ "count": 63433,
229
+ "self": 7.213905162863057,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
+ "total": 135.3240023610499,
233
+ "count": 62567,
234
+ "self": 29.596378664142776,
235
  "children": {
236
  "TorchPolicy.sample_actions": {
237
+ "total": 105.72762369690713,
238
+ "count": 62567,
239
+ "self": 105.72762369690713
240
  }
241
  }
242
  }
243
  }
244
  },
245
  "workers": {
246
+ "total": 1.4456496809802957,
247
+ "count": 63433,
248
  "self": 0.0,
249
  "children": {
250
  "worker_root": {
251
+ "total": 3076.388358235013,
252
+ "count": 63433,
253
  "is_parallel": true,
254
+ "self": 1586.414721414057,
255
  "children": {
256
  "run_training.setup": {
257
  "total": 0.0,
 
260
  "self": 0.0,
261
  "children": {
262
  "steps_from_proto": {
263
+ "total": 0.009083742000029815,
264
  "count": 1,
265
  "is_parallel": true,
266
+ "self": 0.0047278530000767205,
267
  "children": {
268
  "_process_rank_one_or_two_observation": {
269
+ "total": 0.004355888999953095,
270
  "count": 8,
271
  "is_parallel": true,
272
+ "self": 0.004355888999953095
273
  }
274
  }
275
  },
276
  "UnityEnvironment.step": {
277
+ "total": 0.07994424399998934,
278
  "count": 1,
279
  "is_parallel": true,
280
+ "self": 0.0007682590000968048,
281
  "children": {
282
  "UnityEnvironment._generate_step_input": {
283
+ "total": 0.0005485279998538317,
284
  "count": 1,
285
  "is_parallel": true,
286
+ "self": 0.0005485279998538317
287
  },
288
  "communicator.exchange": {
289
+ "total": 0.057014517999959935,
290
  "count": 1,
291
  "is_parallel": true,
292
+ "self": 0.057014517999959935
293
  },
294
  "steps_from_proto": {
295
+ "total": 0.021612939000078768,
296
  "count": 1,
297
  "is_parallel": true,
298
+ "self": 0.0006446339996273309,
299
  "children": {
300
  "_process_rank_one_or_two_observation": {
301
+ "total": 0.020968305000451437,
302
  "count": 8,
303
  "is_parallel": true,
304
+ "self": 0.020968305000451437
305
  }
306
  }
307
  }
 
310
  }
311
  },
312
  "UnityEnvironment.step": {
313
+ "total": 1489.973636820956,
314
+ "count": 63432,
315
  "is_parallel": true,
316
+ "self": 40.21890103792134,
317
  "children": {
318
  "UnityEnvironment._generate_step_input": {
319
+ "total": 23.365737592011783,
320
+ "count": 63432,
321
  "is_parallel": true,
322
+ "self": 23.365737592011783
323
  },
324
  "communicator.exchange": {
325
+ "total": 1299.3704599559792,
326
+ "count": 63432,
327
  "is_parallel": true,
328
+ "self": 1299.3704599559792
329
  },
330
  "steps_from_proto": {
331
+ "total": 127.0185382350437,
332
+ "count": 63432,
333
  "is_parallel": true,
334
+ "self": 34.502453151115105,
335
  "children": {
336
  "_process_rank_one_or_two_observation": {
337
+ "total": 92.5160850839286,
338
+ "count": 507456,
339
  "is_parallel": true,
340
+ "self": 92.5160850839286
341
  }
342
  }
343
  }
 
350
  }
351
  },
352
  "trainer_advance": {
353
+ "total": 1264.2014365680484,
354
+ "count": 63433,
355
+ "self": 4.966577621054967,
356
  "children": {
357
  "process_trajectory": {
358
+ "total": 225.00560161999533,
359
+ "count": 63433,
360
+ "self": 224.59825272099602,
361
  "children": {
362
  "RLTrainer._checkpoint": {
363
+ "total": 0.4073488989993166,
364
  "count": 2,
365
+ "self": 0.4073488989993166
366
  }
367
  }
368
  },
369
  "_update_policy": {
370
+ "total": 1034.229257326998,
371
+ "count": 451,
372
+ "self": 284.0283822679801,
373
  "children": {
374
  "TorchPPOOptimizer.update": {
375
+ "total": 750.200875059018,
376
+ "count": 22830,
377
+ "self": 750.200875059018
378
  }
379
  }
380
  }
 
383
  }
384
  },
385
  "trainer_threads": {
386
+ "total": 1.2909995348309167e-06,
387
  "count": 1,
388
+ "self": 1.2909995348309167e-06
389
  },
390
  "TrainerController._save_models": {
391
+ "total": 0.10764926799947716,
392
  "count": 1,
393
+ "self": 0.0027385729990783148,
394
  "children": {
395
  "RLTrainer._checkpoint": {
396
+ "total": 0.10491069500039885,
397
  "count": 1,
398
+ "self": 0.10491069500039885
399
  }
400
  }
401
  }
run_logs/training_status.json CHANGED
@@ -2,40 +2,40 @@
2
  "Pyramids": {
3
  "checkpoints": [
4
  {
5
- "steps": 499879,
6
- "file_path": "results/Pyramids Training/Pyramids/Pyramids-499879.onnx",
7
- "reward": 0.9571999545608249,
8
- "creation_time": 1658673926.2323952,
9
  "auxillary_file_paths": [
10
- "results/Pyramids Training/Pyramids/Pyramids-499879.pt"
11
  ]
12
  },
13
  {
14
- "steps": 999994,
15
- "file_path": "results/Pyramids Training/Pyramids/Pyramids-999994.onnx",
16
- "reward": null,
17
- "creation_time": 1658675090.9099503,
18
  "auxillary_file_paths": [
19
- "results/Pyramids Training/Pyramids/Pyramids-999994.pt"
20
  ]
21
  },
22
  {
23
- "steps": 1252618,
24
- "file_path": "results/Pyramids Training/Pyramids/Pyramids-1252618.onnx",
25
- "reward": null,
26
- "creation_time": 1658675698.49421,
27
  "auxillary_file_paths": [
28
- "results/Pyramids Training/Pyramids/Pyramids-1252618.pt"
29
  ]
30
  }
31
  ],
32
  "final_checkpoint": {
33
- "steps": 1252618,
34
  "file_path": "results/Pyramids Training/Pyramids.onnx",
35
- "reward": null,
36
- "creation_time": 1658675698.49421,
37
  "auxillary_file_paths": [
38
- "results/Pyramids Training/Pyramids/Pyramids-1252618.pt"
39
  ]
40
  }
41
  },
 
2
  "Pyramids": {
3
  "checkpoints": [
4
  {
5
+ "steps": 499875,
6
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-499875.onnx",
7
+ "reward": 1.5825000256299973,
8
+ "creation_time": 1661156694.2841585,
9
  "auxillary_file_paths": [
10
+ "results/Pyramids Training/Pyramids/Pyramids-499875.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 999954,
15
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-999954.onnx",
16
+ "reward": 1.508999988436699,
17
+ "creation_time": 1661158307.0092275,
18
  "auxillary_file_paths": [
19
+ "results/Pyramids Training/Pyramids/Pyramids-999954.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 1000082,
24
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-1000082.onnx",
25
+ "reward": 1.508999988436699,
26
+ "creation_time": 1661158307.1401775,
27
  "auxillary_file_paths": [
28
+ "results/Pyramids Training/Pyramids/Pyramids-1000082.pt"
29
  ]
30
  }
31
  ],
32
  "final_checkpoint": {
33
+ "steps": 1000082,
34
  "file_path": "results/Pyramids Training/Pyramids.onnx",
35
+ "reward": 1.508999988436699,
36
+ "creation_time": 1661158307.1401775,
37
  "auxillary_file_paths": [
38
+ "results/Pyramids Training/Pyramids/Pyramids-1000082.pt"
39
  ]
40
  }
41
  },