triksblade commited on
Commit
6c9b48f
·
verified ·
1 Parent(s): 6f11892

Continued Training

Browse files
README.md CHANGED
@@ -1,35 +1,35 @@
1
- ---
2
- library_name: ml-agents
3
- tags:
4
- - SoccerTwos
5
- - deep-reinforcement-learning
6
- - reinforcement-learning
7
- - ML-Agents-SoccerTwos
8
- ---
9
-
10
- # **poca** Agent playing **SoccerTwos**
11
- This is a trained model of a **poca** agent playing **SoccerTwos**
12
- using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
-
14
- ## Usage (with ML-Agents)
15
- The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
-
17
- We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
- - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
- browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
- - A *longer tutorial* to understand how works ML-Agents:
21
- https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
-
23
- ### Resume the training
24
- ```bash
25
- mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
- ```
27
-
28
- ### Watch your Agent play
29
- You can watch your agent **playing directly in your browser**
30
-
31
- 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
- 2. Step 1: Find your model_id: triksblade/poca-SoccerTwos
33
- 3. Step 2: Select your *.nn /*.onnx file
34
- 4. Click on Watch the agent play 👀
35
 
 
1
+ ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - SoccerTwos
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-SoccerTwos
8
+ ---
9
+
10
+ # **poca** Agent playing **SoccerTwos**
11
+ This is a trained model of a **poca** agent playing **SoccerTwos**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ ## Usage (with ML-Agents)
15
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
+
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+ - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
+ - A *longer tutorial* to understand how works ML-Agents:
21
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
+
23
+ ### Resume the training
24
+ ```bash
25
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
+ ```
27
+
28
+ ### Watch your Agent play
29
+ You can watch your agent **playing directly in your browser**
30
+
31
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
+ 2. Step 1: Find your model_id: triksblade/poca-SoccerTwos
33
+ 3. Step 2: Select your *.nn /*.onnx file
34
+ 4. Click on Watch the agent play 👀
35
 
SoccerTwos.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d179e28a496f0bdb8e8c95a4ae7faa490e7dace79c8473fbfa09a8cbc03a1ad
3
  size 1768747
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43a583bf9b768b1527bceaf6c7884ff8fc607dee6559683b61424884be28770a
3
  size 1768747
SoccerTwos/SoccerTwos-3499906.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9ba156dc8726c53d580494e81fa1ae45b71d71c7d57fd3664619811dfbf34f5
3
+ size 1768747
SoccerTwos/SoccerTwos-3499906.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d31364e6ec59245e2465f9ed6918d98de6f7b055c0edb812952aba12567a6ab0
3
+ size 28430358
SoccerTwos/SoccerTwos-3999859.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebd83e2d097eb84b17ba8c0a856484bdb39f0b9271d8f31adebecff178a01304
3
+ size 1768747
SoccerTwos/SoccerTwos-3999859.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9ed477c0c33babff90b136efd7d15f99e618545343081f8666b4e7022b3e06d
3
+ size 28430358
SoccerTwos/SoccerTwos-4499974.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c693a625a52b0a203a87d261ef4cfda83a0c7d3779ccf073d67e4fc944c868b6
3
+ size 1768747
SoccerTwos/SoccerTwos-4499974.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd14b7327c7c1d85ced48d4e92dd56ab0061b3926221e3e265fb9ec1f86f1924
3
+ size 28430358
SoccerTwos/SoccerTwos-4999972.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:143de6f8aa2494e97a505635d456bb327c1ecd853c1d54de199caed7d403aa4f
3
+ size 1768747
SoccerTwos/SoccerTwos-4999972.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d1eecd5b4cab7e36856b31aa565cf6e0b4b7557229edec5a18f5569ddcb008c
3
+ size 28430358
SoccerTwos/SoccerTwos-5144723.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43a583bf9b768b1527bceaf6c7884ff8fc607dee6559683b61424884be28770a
3
+ size 1768747
SoccerTwos/SoccerTwos-5144723.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5693d4e7b298823d70c3934a6f9f5c7ef840736d270aa38a578f19e085a8339b
3
+ size 28430358
SoccerTwos/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f10aacc1d9f27faf23101be3285cf619bd5f61cf5c9bbfea4a4dd02e5cc008e
3
  size 28429366
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a88303bb1d4f6513ec41de1dce53473f75860870b62c4a8557bdfc5b11b249c3
3
  size 28429366
SoccerTwos/events.out.tfevents.1716434628.SERVICE-MOBILE.15244.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:601d5287ac6967f5d3cbaf8bbe6188557ed49e954a29873525ff839522825e33
3
+ size 1778267
SoccerTwos/events.out.tfevents.1716449422.SERVICE-MOBILE.5228.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bcecd5fa324686e7a4f0bf6a2dbb912ab2d650505dcf183d761fd16c99eed66
3
+ size 175879
SoccerTwos/events.out.tfevents.1716455014.SERVICE-MOBILE.4032.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dacd19312956f90880f54997cee808c90b09dda0580e90db44ed0ec773da95d
3
+ size 1758423
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SoccerTwos": {"trainer_type": "poca", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "constant", "beta_schedule": "constant", "epsilon_schedule": "constant"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 50000000, "time_horizon": 1000, "summary_freq": 10000, "threaded": false, "self_play": {"save_steps": 50000, "team_change": 200000, "swap_steps": 2000, "window": 10, "play_against_latest_model_ratio": 0.5, "initial_elo": 1200.0}, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/SoccerTwos.exe", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SoccerTwos", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SoccerTwos": {"trainer_type": "poca", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "constant", "beta_schedule": "constant", "epsilon_schedule": "constant"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 50000000, "time_horizon": 1000, "summary_freq": 10000, "threaded": false, "self_play": {"save_steps": 50000, "team_change": 200000, "swap_steps": 2000, "window": 10, "play_against_latest_model_ratio": 0.5, "initial_elo": 1200.0}, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/SoccerTwos.exe", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SoccerTwos", "initialize_from": null, "load_model": false, "resume": true, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -74,7 +74,7 @@ checkpoint_settings:
74
  run_id: SoccerTwos
75
  initialize_from: null
76
  load_model: false
77
- resume: false
78
  force: false
79
  train_model: false
80
  inference: false
 
74
  run_id: SoccerTwos
75
  initialize_from: null
76
  load_model: false
77
+ resume: true
78
  force: false
79
  train_model: false
80
  inference: false
run_logs/Player-0.log CHANGED
@@ -9,7 +9,7 @@ NullGfxDevice:
9
  Renderer: Null Device
10
  Vendor: Unity Technologies
11
  Begin MonoManager ReloadAssembly
12
- - Completed reload, in 9.119 seconds
13
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
14
  Microsoft Media Foundation video decoding to texture disabled: graphics device is Null, only Direct3D 11 and Direct3D 12 (only on desktop) are supported for hardware-accelerated video decoding.
15
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -35,266 +35,15 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
35
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
36
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
37
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
38
- UnloadTime: 1.308500 ms
39
  Registered Communicator in Agent.
40
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
- TOO LONG
42
- TOO LONG
43
- TOO LONG
44
- TOO LONG
45
- TOO LONG
46
- TOO LONG
47
- TOO LONG
48
- TOO LONG
49
- TOO LONG
50
- TOO LONG
51
- TOO LONG
52
- TOO LONG
53
- TOO LONG
54
- TOO LONG
55
- TOO LONG
56
- TOO LONG
57
- TOO LONG
58
- TOO LONG
59
- TOO LONG
60
- TOO LONG
61
- TOO LONG
62
- TOO LONG
63
- TOO LONG
64
- TOO LONG
65
- TOO LONG
66
- TOO LONG
67
- TOO LONG
68
- TOO LONG
69
- TOO LONG
70
- TOO LONG
71
- TOO LONG
72
- TOO LONG
73
- TOO LONG
74
- TOO LONG
75
- TOO LONG
76
- TOO LONG
77
- TOO LONG
78
- TOO LONG
79
- TOO LONG
80
- TOO LONG
81
- TOO LONG
82
- TOO LONG
83
- TOO LONG
84
- TOO LONG
85
- TOO LONG
86
- TOO LONG
87
- TOO LONG
88
- TOO LONG
89
- TOO LONG
90
- TOO LONG
91
- TOO LONG
92
- TOO LONG
93
- TOO LONG
94
- TOO LONG
95
- TOO LONG
96
- TOO LONG
97
- TOO LONG
98
- TOO LONG
99
- TOO LONG
100
- TOO LONG
101
- TOO LONG
102
- TOO LONG
103
- TOO LONG
104
- TOO LONG
105
- TOO LONG
106
- TOO LONG
107
- TOO LONG
108
- TOO LONG
109
- TOO LONG
110
- TOO LONG
111
- TOO LONG
112
- TOO LONG
113
- TOO LONG
114
- TOO LONG
115
- TOO LONG
116
- TOO LONG
117
- TOO LONG
118
- TOO LONG
119
- TOO LONG
120
- TOO LONG
121
- TOO LONG
122
- TOO LONG
123
- TOO LONG
124
- TOO LONG
125
- TOO LONG
126
- TOO LONG
127
- TOO LONG
128
- TOO LONG
129
- TOO LONG
130
- TOO LONG
131
- TOO LONG
132
- TOO LONG
133
- TOO LONG
134
- TOO LONG
135
- TOO LONG
136
- TOO LONG
137
- TOO LONG
138
- TOO LONG
139
- TOO LONG
140
- TOO LONG
141
- TOO LONG
142
- TOO LONG
143
- TOO LONG
144
- TOO LONG
145
- TOO LONG
146
- TOO LONG
147
- TOO LONG
148
- TOO LONG
149
- TOO LONG
150
- TOO LONG
151
- TOO LONG
152
- TOO LONG
153
- TOO LONG
154
- TOO LONG
155
- TOO LONG
156
- TOO LONG
157
- TOO LONG
158
- TOO LONG
159
- TOO LONG
160
- TOO LONG
161
- TOO LONG
162
- TOO LONG
163
- TOO LONG
164
- TOO LONG
165
- TOO LONG
166
- TOO LONG
167
- TOO LONG
168
- TOO LONG
169
- TOO LONG
170
- TOO LONG
171
- TOO LONG
172
- TOO LONG
173
- TOO LONG
174
- TOO LONG
175
- TOO LONG
176
- TOO LONG
177
- TOO LONG
178
- TOO LONG
179
- TOO LONG
180
- TOO LONG
181
- TOO LONG
182
- TOO LONG
183
- TOO LONG
184
- TOO LONG
185
- TOO LONG
186
- TOO LONG
187
- TOO LONG
188
- TOO LONG
189
- TOO LONG
190
- TOO LONG
191
- TOO LONG
192
- TOO LONG
193
- TOO LONG
194
- TOO LONG
195
- TOO LONG
196
- TOO LONG
197
- TOO LONG
198
- TOO LONG
199
- TOO LONG
200
- TOO LONG
201
- TOO LONG
202
- TOO LONG
203
- TOO LONG
204
- TOO LONG
205
- TOO LONG
206
- TOO LONG
207
- TOO LONG
208
- TOO LONG
209
- TOO LONG
210
- TOO LONG
211
- TOO LONG
212
- TOO LONG
213
- TOO LONG
214
- TOO LONG
215
- TOO LONG
216
- TOO LONG
217
- TOO LONG
218
- TOO LONG
219
- TOO LONG
220
- TOO LONG
221
- TOO LONG
222
- TOO LONG
223
- TOO LONG
224
- TOO LONG
225
- TOO LONG
226
- TOO LONG
227
- TOO LONG
228
- TOO LONG
229
- TOO LONG
230
- TOO LONG
231
- TOO LONG
232
- TOO LONG
233
- TOO LONG
234
- TOO LONG
235
- TOO LONG
236
- TOO LONG
237
- TOO LONG
238
- TOO LONG
239
- TOO LONG
240
- TOO LONG
241
- TOO LONG
242
- TOO LONG
243
- TOO LONG
244
- TOO LONG
245
- TOO LONG
246
- TOO LONG
247
- TOO LONG
248
- TOO LONG
249
- TOO LONG
250
- TOO LONG
251
- TOO LONG
252
- TOO LONG
253
- TOO LONG
254
- TOO LONG
255
- TOO LONG
256
- TOO LONG
257
- TOO LONG
258
- TOO LONG
259
- TOO LONG
260
- TOO LONG
261
- TOO LONG
262
- TOO LONG
263
- TOO LONG
264
- TOO LONG
265
- TOO LONG
266
- TOO LONG
267
- TOO LONG
268
- TOO LONG
269
- TOO LONG
270
- TOO LONG
271
- TOO LONG
272
- TOO LONG
273
- TOO LONG
274
- TOO LONG
275
- TOO LONG
276
- TOO LONG
277
- TOO LONG
278
- TOO LONG
279
- TOO LONG
280
- TOO LONG
281
- TOO LONG
282
- TOO LONG
283
- TOO LONG
284
- TOO LONG
285
- TOO LONG
286
- TOO LONG
287
- TOO LONG
288
- TOO LONG
289
- TOO LONG
290
- TOO LONG
291
- TOO LONG
292
  Setting up 4 worker threads for Enlighten.
293
  Memory Statistics:
294
  [ALLOC_TEMP_TLS] TLS Allocator
295
  StackAllocators :
296
  [ALLOC_TEMP_MAIN]
297
- Peak usage frame count: [8.0 KB-16.0 KB]: 29293 frames, [2.0 MB-4.0 MB]: 1 frames
298
  Initial Block Size 4.0 MB
299
  Current Block Size 4.0 MB
300
  Peak Allocated Bytes 2.0 MB
@@ -317,12 +66,12 @@ Memory Statistics:
317
  [ALLOC_TEMP_Job.Worker 6]
318
  Initial Block Size 256.0 KB
319
  Current Block Size 256.0 KB
320
- Peak Allocated Bytes 0 B
321
  Overflow Count 0
322
  [ALLOC_TEMP_Job.Worker 0]
323
  Initial Block Size 256.0 KB
324
  Current Block Size 256.0 KB
325
- Peak Allocated Bytes 0.8 KB
326
  Overflow Count 0
327
  [ALLOC_TEMP_Background Job.Worker 10]
328
  Initial Block Size 32.0 KB
@@ -446,16 +195,16 @@ Memory Statistics:
446
  Used Block count 1
447
  Peak Allocated bytes 0.9 MB
448
  [ALLOC_DEFAULT_MAIN]
449
- Peak usage frame count: [4.0 MB-8.0 MB]: 29294 frames
450
  Requested Block Size 16.0 MB
451
  Peak Block count 1
452
- Peak Allocated memory 5.0 MB
453
  Peak Large allocation bytes 0 B
454
  [ALLOC_DEFAULT_THREAD]
455
- Peak usage frame count: [16.0 MB-32.0 MB]: 29294 frames
456
  Requested Block Size 16.0 MB
457
  Peak Block count 1
458
- Peak Allocated memory 17.4 MB
459
  Peak Large allocation bytes 16.0 MB
460
  [ALLOC_TEMP_JOB_1_FRAME]
461
  Initial Block Size 2.0 MB
@@ -484,13 +233,13 @@ Memory Statistics:
484
  Used Block count 1
485
  Peak Allocated bytes 0.9 MB
486
  [ALLOC_GFX_MAIN]
487
- Peak usage frame count: [32.0 KB-64.0 KB]: 29293 frames, [64.0 KB-128.0 KB]: 1 frames
488
  Requested Block Size 16.0 MB
489
  Peak Block count 1
490
  Peak Allocated memory 65.6 KB
491
  Peak Large allocation bytes 0 B
492
  [ALLOC_GFX_THREAD]
493
- Peak usage frame count: [128.0 KB-256.0 KB]: 29294 frames
494
  Requested Block Size 16.0 MB
495
  Peak Block count 1
496
  Peak Allocated memory 173.6 KB
@@ -502,16 +251,16 @@ Memory Statistics:
502
  Used Block count 1
503
  Peak Allocated bytes 0.9 MB
504
  [ALLOC_CACHEOBJECTS_MAIN]
505
- Peak usage frame count: [0.5 MB-1.0 MB]: 29294 frames
506
  Requested Block Size 4.0 MB
507
  Peak Block count 1
508
  Peak Allocated memory 0.6 MB
509
  Peak Large allocation bytes 0 B
510
  [ALLOC_CACHEOBJECTS_THREAD]
511
- Peak usage frame count: [0.5 MB-1.0 MB]: 29293 frames, [2.0 MB-4.0 MB]: 1 frames
512
  Requested Block Size 4.0 MB
513
  Peak Block count 1
514
- Peak Allocated memory 2.2 MB
515
  Peak Large allocation bytes 0 B
516
  [ALLOC_TYPETREE] Dual Thread Allocator
517
  Peak main deferred allocation count 0
@@ -520,13 +269,13 @@ Memory Statistics:
520
  Used Block count 1
521
  Peak Allocated bytes 0.9 MB
522
  [ALLOC_TYPETREE_MAIN]
523
- Peak usage frame count: [0-1.0 KB]: 29294 frames
524
  Requested Block Size 2.0 MB
525
  Peak Block count 1
526
  Peak Allocated memory 1.0 KB
527
  Peak Large allocation bytes 0 B
528
  [ALLOC_TYPETREE_THREAD]
529
- Peak usage frame count: [1.0 KB-2.0 KB]: 29293 frames, [2.0 KB-4.0 KB]: 1 frames
530
  Requested Block Size 2.0 MB
531
  Peak Block count 1
532
  Peak Allocated memory 2.2 KB
 
9
  Renderer: Null Device
10
  Vendor: Unity Technologies
11
  Begin MonoManager ReloadAssembly
12
+ - Completed reload, in 8.918 seconds
13
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
14
  Microsoft Media Foundation video decoding to texture disabled: graphics device is Null, only Direct3D 11 and Direct3D 12 (only on desktop) are supported for hardware-accelerated video decoding.
15
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
35
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
36
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
37
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
38
+ UnloadTime: 1.341700 ms
39
  Registered Communicator in Agent.
40
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  Setting up 4 worker threads for Enlighten.
42
  Memory Statistics:
43
  [ALLOC_TEMP_TLS] TLS Allocator
44
  StackAllocators :
45
  [ALLOC_TEMP_MAIN]
46
+ Peak usage frame count: [8.0 KB-16.0 KB]: 32145 frames, [2.0 MB-4.0 MB]: 1 frames
47
  Initial Block Size 4.0 MB
48
  Current Block Size 4.0 MB
49
  Peak Allocated Bytes 2.0 MB
 
66
  [ALLOC_TEMP_Job.Worker 6]
67
  Initial Block Size 256.0 KB
68
  Current Block Size 256.0 KB
69
+ Peak Allocated Bytes 0.8 KB
70
  Overflow Count 0
71
  [ALLOC_TEMP_Job.Worker 0]
72
  Initial Block Size 256.0 KB
73
  Current Block Size 256.0 KB
74
+ Peak Allocated Bytes 0 B
75
  Overflow Count 0
76
  [ALLOC_TEMP_Background Job.Worker 10]
77
  Initial Block Size 32.0 KB
 
195
  Used Block count 1
196
  Peak Allocated bytes 0.9 MB
197
  [ALLOC_DEFAULT_MAIN]
198
+ Peak usage frame count: [4.0 MB-8.0 MB]: 32146 frames
199
  Requested Block Size 16.0 MB
200
  Peak Block count 1
201
+ Peak Allocated memory 4.8 MB
202
  Peak Large allocation bytes 0 B
203
  [ALLOC_DEFAULT_THREAD]
204
+ Peak usage frame count: [16.0 MB-32.0 MB]: 32146 frames
205
  Requested Block Size 16.0 MB
206
  Peak Block count 1
207
+ Peak Allocated memory 17.5 MB
208
  Peak Large allocation bytes 16.0 MB
209
  [ALLOC_TEMP_JOB_1_FRAME]
210
  Initial Block Size 2.0 MB
 
233
  Used Block count 1
234
  Peak Allocated bytes 0.9 MB
235
  [ALLOC_GFX_MAIN]
236
+ Peak usage frame count: [32.0 KB-64.0 KB]: 32145 frames, [64.0 KB-128.0 KB]: 1 frames
237
  Requested Block Size 16.0 MB
238
  Peak Block count 1
239
  Peak Allocated memory 65.6 KB
240
  Peak Large allocation bytes 0 B
241
  [ALLOC_GFX_THREAD]
242
+ Peak usage frame count: [128.0 KB-256.0 KB]: 32146 frames
243
  Requested Block Size 16.0 MB
244
  Peak Block count 1
245
  Peak Allocated memory 173.6 KB
 
251
  Used Block count 1
252
  Peak Allocated bytes 0.9 MB
253
  [ALLOC_CACHEOBJECTS_MAIN]
254
+ Peak usage frame count: [0.5 MB-1.0 MB]: 32146 frames
255
  Requested Block Size 4.0 MB
256
  Peak Block count 1
257
  Peak Allocated memory 0.6 MB
258
  Peak Large allocation bytes 0 B
259
  [ALLOC_CACHEOBJECTS_THREAD]
260
+ Peak usage frame count: [0.5 MB-1.0 MB]: 32145 frames, [2.0 MB-4.0 MB]: 1 frames
261
  Requested Block Size 4.0 MB
262
  Peak Block count 1
263
+ Peak Allocated memory 2.7 MB
264
  Peak Large allocation bytes 0 B
265
  [ALLOC_TYPETREE] Dual Thread Allocator
266
  Peak main deferred allocation count 0
 
269
  Used Block count 1
270
  Peak Allocated bytes 0.9 MB
271
  [ALLOC_TYPETREE_MAIN]
272
+ Peak usage frame count: [0-1.0 KB]: 32146 frames
273
  Requested Block Size 2.0 MB
274
  Peak Block count 1
275
  Peak Allocated memory 1.0 KB
276
  Peak Large allocation bytes 0 B
277
  [ALLOC_TYPETREE_THREAD]
278
+ Peak usage frame count: [1.0 KB-2.0 KB]: 32145 frames, [2.0 KB-4.0 KB]: 1 frames
279
  Requested Block Size 2.0 MB
280
  Peak Block count 1
281
  Peak Allocated memory 2.2 KB
run_logs/timers.json CHANGED
@@ -2,305 +2,305 @@
2
  "name": "root",
3
  "gauges": {
4
  "SoccerTwos.Policy.Entropy.mean": {
5
- "value": 2.60512113571167,
6
- "min": 2.576058864593506,
7
- "max": 3.2957584857940674,
8
- "count": 153
9
  },
10
  "SoccerTwos.Policy.Entropy.sum": {
11
- "value": 49768.234375,
12
- "min": 19701.3984375,
13
- "max": 105464.1875,
14
- "count": 153
15
  },
16
  "SoccerTwos.Environment.EpisodeLength.mean": {
17
- "value": 70.44285714285714,
18
- "min": 55.69318181818182,
19
- "max": 999.0,
20
- "count": 153
21
  },
22
  "SoccerTwos.Environment.EpisodeLength.sum": {
23
- "value": 19724.0,
24
- "min": 16380.0,
25
- "max": 25536.0,
26
- "count": 153
27
  },
28
  "SoccerTwos.Self-play.ELO.mean": {
29
- "value": 1432.0507427110429,
30
- "min": 1198.9900907842039,
31
- "max": 1439.0801697786028,
32
- "count": 152
33
  },
34
  "SoccerTwos.Self-play.ELO.sum": {
35
- "value": 200487.103979546,
36
- "min": 2397.9801815684077,
37
- "max": 245331.37821920833,
38
- "count": 152
39
  },
40
  "SoccerTwos.Step.mean": {
41
- "value": 1529960.0,
42
- "min": 9786.0,
43
- "max": 1529960.0,
44
- "count": 153
45
  },
46
  "SoccerTwos.Step.sum": {
47
- "value": 1529960.0,
48
- "min": 9786.0,
49
- "max": 1529960.0,
50
- "count": 153
51
  },
52
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
53
- "value": 0.08859968930482864,
54
- "min": -0.03122873045504093,
55
- "max": 0.25784891843795776,
56
- "count": 153
57
  },
58
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
59
- "value": 12.315357208251953,
60
- "min": -4.340793609619141,
61
- "max": 26.724498748779297,
62
- "count": 153
63
  },
64
  "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
65
- "value": 0.07361045479774475,
66
- "min": -0.03681948408484459,
67
- "max": 0.25843432545661926,
68
- "count": 153
69
  },
70
  "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
71
- "value": 10.231853485107422,
72
- "min": -5.117908477783203,
73
- "max": 25.840808868408203,
74
- "count": 153
75
  },
76
  "SoccerTwos.Environment.CumulativeReward.mean": {
77
  "value": 0.0,
78
  "min": 0.0,
79
  "max": 0.0,
80
- "count": 153
81
  },
82
  "SoccerTwos.Environment.CumulativeReward.sum": {
83
  "value": 0.0,
84
  "min": 0.0,
85
  "max": 0.0,
86
- "count": 153
87
  },
88
  "SoccerTwos.Policy.ExtrinsicReward.mean": {
89
- "value": 0.12326043329650549,
90
- "min": -0.5714285714285714,
91
- "max": 0.6532827560929046,
92
- "count": 153
93
  },
94
  "SoccerTwos.Policy.ExtrinsicReward.sum": {
95
- "value": 17.133200228214264,
96
- "min": -21.79960000514984,
97
- "max": 56.8355997800827,
98
- "count": 153
99
  },
100
  "SoccerTwos.Environment.GroupCumulativeReward.mean": {
101
- "value": 0.12326043329650549,
102
- "min": -0.5714285714285714,
103
- "max": 0.6532827560929046,
104
- "count": 153
105
  },
106
  "SoccerTwos.Environment.GroupCumulativeReward.sum": {
107
- "value": 17.133200228214264,
108
- "min": -21.79960000514984,
109
- "max": 56.8355997800827,
110
- "count": 153
111
  },
112
  "SoccerTwos.IsTraining.mean": {
113
  "value": 1.0,
114
  "min": 1.0,
115
  "max": 1.0,
116
- "count": 153
117
  },
118
  "SoccerTwos.IsTraining.sum": {
119
  "value": 1.0,
120
  "min": 1.0,
121
  "max": 1.0,
122
- "count": 153
123
  },
124
  "SoccerTwos.Losses.PolicyLoss.mean": {
125
- "value": 0.017969103282666765,
126
- "min": 0.010487537668571653,
127
- "max": 0.021603452736356606,
128
- "count": 72
129
  },
130
  "SoccerTwos.Losses.PolicyLoss.sum": {
131
- "value": 0.017969103282666765,
132
- "min": 0.010487537668571653,
133
- "max": 0.021603452736356606,
134
- "count": 72
135
  },
136
  "SoccerTwos.Losses.ValueLoss.mean": {
137
- "value": 0.08188363139828046,
138
- "min": 0.0016951362524802486,
139
- "max": 0.08501399084925651,
140
- "count": 72
141
  },
142
  "SoccerTwos.Losses.ValueLoss.sum": {
143
- "value": 0.08188363139828046,
144
- "min": 0.0016951362524802486,
145
- "max": 0.08501399084925651,
146
- "count": 72
147
  },
148
  "SoccerTwos.Losses.BaselineLoss.mean": {
149
- "value": 0.08477470974127452,
150
- "min": 0.0017042422861171266,
151
- "max": 0.08773589010039966,
152
- "count": 72
153
  },
154
  "SoccerTwos.Losses.BaselineLoss.sum": {
155
- "value": 0.08477470974127452,
156
- "min": 0.0017042422861171266,
157
- "max": 0.08773589010039966,
158
- "count": 72
159
  },
160
  "SoccerTwos.Policy.LearningRate.mean": {
161
  "value": 0.0003,
162
  "min": 0.0003,
163
  "max": 0.0003,
164
- "count": 72
165
  },
166
  "SoccerTwos.Policy.LearningRate.sum": {
167
  "value": 0.0003,
168
  "min": 0.0003,
169
  "max": 0.0003,
170
- "count": 72
171
  },
172
  "SoccerTwos.Policy.Epsilon.mean": {
173
  "value": 0.20000000000000007,
174
  "min": 0.20000000000000007,
175
  "max": 0.20000000000000007,
176
- "count": 72
177
  },
178
  "SoccerTwos.Policy.Epsilon.sum": {
179
  "value": 0.20000000000000007,
180
  "min": 0.20000000000000007,
181
  "max": 0.20000000000000007,
182
- "count": 72
183
  },
184
  "SoccerTwos.Policy.Beta.mean": {
185
  "value": 0.005000000000000001,
186
  "min": 0.005000000000000001,
187
  "max": 0.005000000000000001,
188
- "count": 72
189
  },
190
  "SoccerTwos.Policy.Beta.sum": {
191
  "value": 0.005000000000000001,
192
  "min": 0.005000000000000001,
193
  "max": 0.005000000000000001,
194
- "count": 72
195
  }
196
  },
197
  "metadata": {
198
  "timer_format_version": "0.1.0",
199
- "start_time_seconds": "1716284271",
200
  "python_version": "3.10.12 | packaged by Anaconda, Inc. | (main, Jul 5 2023, 19:01:18) [MSC v.1916 64 bit (AMD64)]",
201
- "command_line_arguments": "\\\\?\\C:\\Users\\ASUSROG\\.conda\\envs\\rl\\Scripts\\mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos --no-graphics",
202
  "mlagents_version": "1.1.0.dev0",
203
  "mlagents_envs_version": "1.1.0.dev0",
204
  "communication_protocol_version": "1.5.0",
205
  "pytorch_version": "2.3.0+cpu",
206
  "numpy_version": "1.23.5",
207
- "end_time_seconds": "1716295092"
208
  },
209
- "total": 10820.996979999938,
210
  "count": 1,
211
- "self": 1.0817919998662546,
212
  "children": {
213
  "run_training.setup": {
214
- "total": 0.2845190000953153,
215
  "count": 1,
216
- "self": 0.2845190000953153
217
  },
218
  "TrainerController.start_learning": {
219
- "total": 10819.630668999976,
220
  "count": 1,
221
- "self": 7.065667584189214,
222
  "children": {
223
  "TrainerController._reset_env": {
224
- "total": 21.102362099918537,
225
- "count": 8,
226
- "self": 21.102362099918537
227
  },
228
  "TrainerController.advance": {
229
- "total": 10791.088720715838,
230
- "count": 101527,
231
- "self": 7.374858103692532,
232
  "children": {
233
  "env_step": {
234
- "total": 5743.239232376101,
235
- "count": 101527,
236
- "self": 4415.3459142574575,
237
  "children": {
238
  "SubprocessEnvManager._take_step": {
239
- "total": 1323.189395701862,
240
- "count": 101527,
241
- "self": 43.19361850025598,
242
  "children": {
243
  "TorchPolicy.evaluate": {
244
- "total": 1279.995777201606,
245
- "count": 195294,
246
- "self": 1279.995777201606
247
  }
248
  }
249
  },
250
  "workers": {
251
- "total": 4.703922416782007,
252
- "count": 101527,
253
  "self": 0.0,
254
  "children": {
255
  "worker_root": {
256
- "total": 10749.106660613208,
257
- "count": 101527,
258
  "is_parallel": true,
259
- "self": 7286.344790431438,
260
  "children": {
261
  "steps_from_proto": {
262
- "total": 0.04099139990285039,
263
- "count": 16,
264
  "is_parallel": true,
265
- "self": 0.00800789927598089,
266
  "children": {
267
  "_process_rank_one_or_two_observation": {
268
- "total": 0.0329835006268695,
269
- "count": 64,
270
  "is_parallel": true,
271
- "self": 0.0329835006268695
272
  }
273
  }
274
  },
275
  "UnityEnvironment.step": {
276
- "total": 3462.720878781867,
277
- "count": 101527,
278
  "is_parallel": true,
279
- "self": 200.90259891434107,
280
  "children": {
281
  "UnityEnvironment._generate_step_input": {
282
- "total": 198.3468875954859,
283
- "count": 101527,
284
  "is_parallel": true,
285
- "self": 198.3468875954859
286
  },
287
  "communicator.exchange": {
288
- "total": 2370.8362009989796,
289
- "count": 101527,
290
  "is_parallel": true,
291
- "self": 2370.8362009989796
292
  },
293
  "steps_from_proto": {
294
- "total": 692.6351912730606,
295
- "count": 203054,
296
  "is_parallel": true,
297
- "self": 133.0821683453396,
298
  "children": {
299
  "_process_rank_one_or_two_observation": {
300
- "total": 559.553022927721,
301
- "count": 812216,
302
  "is_parallel": true,
303
- "self": 559.553022927721
304
  }
305
  }
306
  }
@@ -313,31 +313,31 @@
313
  }
314
  },
315
  "trainer_advance": {
316
- "total": 5040.474630236044,
317
- "count": 101527,
318
- "self": 41.68894123204518,
319
  "children": {
320
  "process_trajectory": {
321
- "total": 1336.936692404328,
322
- "count": 101527,
323
- "self": 1335.9088755045086,
324
  "children": {
325
  "RLTrainer._checkpoint": {
326
- "total": 1.0278168998192996,
327
- "count": 3,
328
- "self": 1.0278168998192996
329
  }
330
  }
331
  },
332
  "_update_policy": {
333
- "total": 3661.8489965996705,
334
- "count": 73,
335
- "self": 485.0931242010556,
336
  "children": {
337
  "TorchPOCAOptimizer.update": {
338
- "total": 3176.755872398615,
339
- "count": 2186,
340
- "self": 3176.755872398615
341
  }
342
  }
343
  }
@@ -346,19 +346,19 @@
346
  }
347
  },
348
  "trainer_threads": {
349
- "total": 3.100023604929447e-06,
350
  "count": 1,
351
- "self": 3.100023604929447e-06
352
  },
353
  "TrainerController._save_models": {
354
- "total": 0.3739155000075698,
355
  "count": 1,
356
- "self": 0.016544800018891692,
357
  "children": {
358
  "RLTrainer._checkpoint": {
359
- "total": 0.3573706999886781,
360
  "count": 1,
361
- "self": 0.3573706999886781
362
  }
363
  }
364
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SoccerTwos.Policy.Entropy.mean": {
5
+ "value": 1.9214180707931519,
6
+ "min": 1.862352967262268,
7
+ "max": 2.0375585556030273,
8
+ "count": 171
9
  },
10
  "SoccerTwos.Policy.Entropy.sum": {
11
+ "value": 39473.61328125,
12
+ "min": 14372.5556640625,
13
+ "max": 44608.6640625,
14
+ "count": 171
15
  },
16
  "SoccerTwos.Environment.EpisodeLength.mean": {
17
+ "value": 49.58585858585859,
18
+ "min": 40.706896551724135,
19
+ "max": 71.81428571428572,
20
+ "count": 171
21
  },
22
  "SoccerTwos.Environment.EpisodeLength.sum": {
23
+ "value": 19636.0,
24
+ "min": 6180.0,
25
+ "max": 20260.0,
26
+ "count": 171
27
  },
28
  "SoccerTwos.Self-play.ELO.mean": {
29
+ "value": 1585.7677991109404,
30
+ "min": 1553.5179580138852,
31
+ "max": 1610.7972019352037,
32
+ "count": 171
33
  },
34
  "SoccerTwos.Self-play.ELO.sum": {
35
+ "value": 313982.0242239662,
36
+ "min": 94213.86956426132,
37
+ "max": 394536.63157583587,
38
+ "count": 171
39
  },
40
  "SoccerTwos.Step.mean": {
41
+ "value": 5139959.0,
42
+ "min": 3439951.0,
43
+ "max": 5139959.0,
44
+ "count": 171
45
  },
46
  "SoccerTwos.Step.sum": {
47
+ "value": 5139959.0,
48
+ "min": 3439951.0,
49
+ "max": 5139959.0,
50
+ "count": 171
51
  },
52
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
53
+ "value": -0.060572218149900436,
54
+ "min": -0.08898256719112396,
55
+ "max": 0.1005156859755516,
56
+ "count": 171
57
  },
58
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
59
+ "value": -12.053871154785156,
60
+ "min": -17.240734100341797,
61
+ "max": 17.89179229736328,
62
+ "count": 171
63
  },
64
  "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
65
+ "value": -0.05699760094285011,
66
+ "min": -0.087178535759449,
67
+ "max": 0.09999766200780869,
68
+ "count": 171
69
  },
70
  "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
71
+ "value": -11.342522621154785,
72
+ "min": -16.390785217285156,
73
+ "max": 17.799583435058594,
74
+ "count": 171
75
  },
76
  "SoccerTwos.Environment.CumulativeReward.mean": {
77
  "value": 0.0,
78
  "min": 0.0,
79
  "max": 0.0,
80
+ "count": 171
81
  },
82
  "SoccerTwos.Environment.CumulativeReward.sum": {
83
  "value": 0.0,
84
  "min": 0.0,
85
  "max": 0.0,
86
+ "count": 171
87
  },
88
  "SoccerTwos.Policy.ExtrinsicReward.mean": {
89
+ "value": -0.1673165817356589,
90
+ "min": -0.2861095484177671,
91
+ "max": 0.22272921378692884,
92
+ "count": 171
93
  },
94
  "SoccerTwos.Policy.ExtrinsicReward.sum": {
95
+ "value": -33.29599976539612,
96
+ "min": -56.93580013513565,
97
+ "max": 39.645800054073334,
98
+ "count": 171
99
  },
100
  "SoccerTwos.Environment.GroupCumulativeReward.mean": {
101
+ "value": -0.1673165817356589,
102
+ "min": -0.2861095484177671,
103
+ "max": 0.22272921378692884,
104
+ "count": 171
105
  },
106
  "SoccerTwos.Environment.GroupCumulativeReward.sum": {
107
+ "value": -33.29599976539612,
108
+ "min": -56.93580013513565,
109
+ "max": 39.645800054073334,
110
+ "count": 171
111
  },
112
  "SoccerTwos.IsTraining.mean": {
113
  "value": 1.0,
114
  "min": 1.0,
115
  "max": 1.0,
116
+ "count": 171
117
  },
118
  "SoccerTwos.IsTraining.sum": {
119
  "value": 1.0,
120
  "min": 1.0,
121
  "max": 1.0,
122
+ "count": 171
123
  },
124
  "SoccerTwos.Losses.PolicyLoss.mean": {
125
+ "value": 0.018759650674959025,
126
+ "min": 0.01115469560221148,
127
+ "max": 0.021448074255992346,
128
+ "count": 82
129
  },
130
  "SoccerTwos.Losses.PolicyLoss.sum": {
131
+ "value": 0.018759650674959025,
132
+ "min": 0.01115469560221148,
133
+ "max": 0.021448074255992346,
134
+ "count": 82
135
  },
136
  "SoccerTwos.Losses.ValueLoss.mean": {
137
+ "value": 0.11596904620528221,
138
+ "min": 0.09423803612589836,
139
+ "max": 0.12632074877619742,
140
+ "count": 82
141
  },
142
  "SoccerTwos.Losses.ValueLoss.sum": {
143
+ "value": 0.11596904620528221,
144
+ "min": 0.09423803612589836,
145
+ "max": 0.12632074877619742,
146
+ "count": 82
147
  },
148
  "SoccerTwos.Losses.BaselineLoss.mean": {
149
+ "value": 0.11812345335880915,
150
+ "min": 0.0952321524421374,
151
+ "max": 0.12842872540156047,
152
+ "count": 82
153
  },
154
  "SoccerTwos.Losses.BaselineLoss.sum": {
155
+ "value": 0.11812345335880915,
156
+ "min": 0.0952321524421374,
157
+ "max": 0.12842872540156047,
158
+ "count": 82
159
  },
160
  "SoccerTwos.Policy.LearningRate.mean": {
161
  "value": 0.0003,
162
  "min": 0.0003,
163
  "max": 0.0003,
164
+ "count": 82
165
  },
166
  "SoccerTwos.Policy.LearningRate.sum": {
167
  "value": 0.0003,
168
  "min": 0.0003,
169
  "max": 0.0003,
170
+ "count": 82
171
  },
172
  "SoccerTwos.Policy.Epsilon.mean": {
173
  "value": 0.20000000000000007,
174
  "min": 0.20000000000000007,
175
  "max": 0.20000000000000007,
176
+ "count": 82
177
  },
178
  "SoccerTwos.Policy.Epsilon.sum": {
179
  "value": 0.20000000000000007,
180
  "min": 0.20000000000000007,
181
  "max": 0.20000000000000007,
182
+ "count": 82
183
  },
184
  "SoccerTwos.Policy.Beta.mean": {
185
  "value": 0.005000000000000001,
186
  "min": 0.005000000000000001,
187
  "max": 0.005000000000000001,
188
+ "count": 82
189
  },
190
  "SoccerTwos.Policy.Beta.sum": {
191
  "value": 0.005000000000000001,
192
  "min": 0.005000000000000001,
193
  "max": 0.005000000000000001,
194
+ "count": 82
195
  }
196
  },
197
  "metadata": {
198
  "timer_format_version": "0.1.0",
199
+ "start_time_seconds": "1716454994",
200
  "python_version": "3.10.12 | packaged by Anaconda, Inc. | (main, Jul 5 2023, 19:01:18) [MSC v.1916 64 bit (AMD64)]",
201
+ "command_line_arguments": "\\\\?\\C:\\Users\\ASUSROG\\.conda\\envs\\rl\\Scripts\\mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos --no-graphics --resume",
202
  "mlagents_version": "1.1.0.dev0",
203
  "mlagents_envs_version": "1.1.0.dev0",
204
  "communication_protocol_version": "1.5.0",
205
  "pytorch_version": "2.3.0+cpu",
206
  "numpy_version": "1.23.5",
207
+ "end_time_seconds": "1716467265"
208
  },
209
+ "total": 12270.348834600001,
210
  "count": 1,
211
+ "self": 1.1021543000024394,
212
  "children": {
213
  "run_training.setup": {
214
+ "total": 0.3115802999996049,
215
  "count": 1,
216
+ "self": 0.3115802999996049
217
  },
218
  "TrainerController.start_learning": {
219
+ "total": 12268.935099999999,
220
  "count": 1,
221
+ "self": 8.741336000073716,
222
  "children": {
223
  "TrainerController._reset_env": {
224
+ "total": 24.339133400000264,
225
+ "count": 10,
226
+ "self": 24.339133400000264
227
  },
228
  "TrainerController.advance": {
229
+ "total": 12235.582564599925,
230
+ "count": 119838,
231
+ "self": 8.518548199779616,
232
  "children": {
233
  "env_step": {
234
+ "total": 6489.675793600161,
235
+ "count": 119838,
236
+ "self": 5023.78484800052,
237
  "children": {
238
  "SubprocessEnvManager._take_step": {
239
+ "total": 1460.3441352998884,
240
+ "count": 119838,
241
+ "self": 46.86335090045304,
242
  "children": {
243
  "TorchPolicy.evaluate": {
244
+ "total": 1413.4807843994354,
245
+ "count": 214308,
246
+ "self": 1413.4807843994354
247
  }
248
  }
249
  },
250
  "workers": {
251
+ "total": 5.546810299753361,
252
+ "count": 119838,
253
  "self": 0.0,
254
  "children": {
255
  "worker_root": {
256
+ "total": 12204.119910299984,
257
+ "count": 119838,
258
  "is_parallel": true,
259
+ "self": 8252.098853600017,
260
  "children": {
261
  "steps_from_proto": {
262
+ "total": 0.0566629999966608,
263
+ "count": 20,
264
  "is_parallel": true,
265
+ "self": 0.011537199997746939,
266
  "children": {
267
  "_process_rank_one_or_two_observation": {
268
+ "total": 0.04512579999891386,
269
+ "count": 80,
270
  "is_parallel": true,
271
+ "self": 0.04512579999891386
272
  }
273
  }
274
  },
275
  "UnityEnvironment.step": {
276
+ "total": 3951.9643936999705,
277
+ "count": 119838,
278
  "is_parallel": true,
279
+ "self": 227.40118399990342,
280
  "children": {
281
  "UnityEnvironment._generate_step_input": {
282
+ "total": 221.29594099971882,
283
+ "count": 119838,
284
  "is_parallel": true,
285
+ "self": 221.29594099971882
286
  },
287
  "communicator.exchange": {
288
+ "total": 2718.435361500345,
289
+ "count": 119838,
290
  "is_parallel": true,
291
+ "self": 2718.435361500345
292
  },
293
  "steps_from_proto": {
294
+ "total": 784.8319072000031,
295
+ "count": 239676,
296
  "is_parallel": true,
297
+ "self": 154.5026458000434,
298
  "children": {
299
  "_process_rank_one_or_two_observation": {
300
+ "total": 630.3292613999597,
301
+ "count": 958704,
302
  "is_parallel": true,
303
+ "self": 630.3292613999597
304
  }
305
  }
306
  }
 
313
  }
314
  },
315
  "trainer_advance": {
316
+ "total": 5737.388222799984,
317
+ "count": 119838,
318
+ "self": 47.69191030006823,
319
  "children": {
320
  "process_trajectory": {
321
+ "total": 1710.6069001999222,
322
+ "count": 119838,
323
+ "self": 1709.2757095999232,
324
  "children": {
325
  "RLTrainer._checkpoint": {
326
+ "total": 1.3311905999989904,
327
+ "count": 4,
328
+ "self": 1.3311905999989904
329
  }
330
  }
331
  },
332
  "_update_policy": {
333
+ "total": 3979.0894122999935,
334
+ "count": 83,
335
+ "self": 503.9244860999961,
336
  "children": {
337
  "TorchPOCAOptimizer.update": {
338
+ "total": 3475.1649261999974,
339
+ "count": 2480,
340
+ "self": 3475.1649261999974
341
  }
342
  }
343
  }
 
346
  }
347
  },
348
  "trainer_threads": {
349
+ "total": 2.599999788799323e-06,
350
  "count": 1,
351
+ "self": 2.599999788799323e-06
352
  },
353
  "TrainerController._save_models": {
354
+ "total": 0.2720633999997517,
355
  "count": 1,
356
+ "self": 0.014902599999913946,
357
  "children": {
358
  "RLTrainer._checkpoint": {
359
+ "total": 0.25716079999983776,
360
  "count": 1,
361
+ "self": 0.25716079999983776
362
  }
363
  }
364
  }
run_logs/training_status.json CHANGED
@@ -2,50 +2,59 @@
2
  "SoccerTwos": {
3
  "checkpoints": [
4
  {
5
- "steps": 499442,
6
- "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-499442.onnx",
7
  "reward": 0.0,
8
- "creation_time": 1716287975.3044803,
9
  "auxillary_file_paths": [
10
- "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-499442.pt"
11
  ]
12
  },
13
  {
14
- "steps": 999768,
15
- "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-999768.onnx",
16
  "reward": 0.0,
17
- "creation_time": 1716291286.4634547,
18
  "auxillary_file_paths": [
19
- "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-999768.pt"
20
  ]
21
  },
22
  {
23
- "steps": 1499994,
24
- "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-1499994.onnx",
25
  "reward": 0.0,
26
- "creation_time": 1716294816.3197072,
27
  "auxillary_file_paths": [
28
- "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-1499994.pt"
29
  ]
30
  },
31
  {
32
- "steps": 1534236,
33
- "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-1534236.onnx",
 
 
 
 
 
 
 
 
 
34
  "reward": null,
35
- "creation_time": 1716295091.7712605,
36
  "auxillary_file_paths": [
37
- "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-1534236.pt"
38
  ]
39
  }
40
  ],
41
- "elo": 1431.976231608046,
42
  "final_checkpoint": {
43
- "steps": 1534236,
44
  "file_path": "results\\SoccerTwos\\SoccerTwos.onnx",
45
  "reward": null,
46
- "creation_time": 1716295091.7712605,
47
  "auxillary_file_paths": [
48
- "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-1534236.pt"
49
  ]
50
  }
51
  },
 
2
  "SoccerTwos": {
3
  "checkpoints": [
4
  {
5
+ "steps": 3499906,
6
+ "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-3499906.onnx",
7
  "reward": 0.0,
8
+ "creation_time": 1716455525.8830073,
9
  "auxillary_file_paths": [
10
+ "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-3499906.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 3999859,
15
+ "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-3999859.onnx",
16
  "reward": 0.0,
17
+ "creation_time": 1716459059.8025615,
18
  "auxillary_file_paths": [
19
+ "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-3999859.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 4499974,
24
+ "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-4499974.onnx",
25
  "reward": 0.0,
26
+ "creation_time": 1716462597.8991349,
27
  "auxillary_file_paths": [
28
+ "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-4499974.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 4999972,
33
+ "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-4999972.onnx",
34
+ "reward": 0.0,
35
+ "creation_time": 1716466198.6816692,
36
+ "auxillary_file_paths": [
37
+ "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-4999972.pt"
38
+ ]
39
+ },
40
+ {
41
+ "steps": 5144723,
42
+ "file_path": "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-5144723.onnx",
43
  "reward": null,
44
+ "creation_time": 1716467264.04394,
45
  "auxillary_file_paths": [
46
+ "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-5144723.pt"
47
  ]
48
  }
49
  ],
50
+ "elo": 1585.1402033727863,
51
  "final_checkpoint": {
52
+ "steps": 5144723,
53
  "file_path": "results\\SoccerTwos\\SoccerTwos.onnx",
54
  "reward": null,
55
+ "creation_time": 1716467264.04394,
56
  "auxillary_file_paths": [
57
+ "results\\SoccerTwos\\SoccerTwos\\SoccerTwos-5144723.pt"
58
  ]
59
  }
60
  },