giovannidispoto commited on
Commit
4b62b3d
·
1 Parent(s): ceec6cc

First Push`

Browse files
SoccerTwos.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dfc2027c8dc413744c2583346f9f67cc0b958d963f7c3780bc48720f85d6d0e
3
  size 1764633
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70506f8cee310950c62ab165bf7ebe724d52a1128a883c02b70f912fa671cdec
3
  size 1764633
SoccerTwos/SoccerTwos-2606308.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70506f8cee310950c62ab165bf7ebe724d52a1128a883c02b70f912fa671cdec
3
+ size 1764633
SoccerTwos/SoccerTwos-2606308.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15283d56cd77101e66587ac8b1b8c709476a8bc1d03bc14efa34fec302849a1a
3
+ size 28421201
SoccerTwos/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b7e741047e393067f5b7a9f490918afdd2ef3a3fa37520f35ec4886eee61064
3
  size 28421201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15283d56cd77101e66587ac8b1b8c709476a8bc1d03bc14efa34fec302849a1a
3
  size 28421201
SoccerTwos/events.out.tfevents.1689924789.MBP-di-Giovanni.homenet.telecomitalia.it.6897.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:311cb6f1c49d3b88ef76ccfba8bd256bc62009f2ba7fba6aa894845f35019004
3
- size 1393000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20732f2193855ef01e0b403a4289fb858ad94aa8b253245668de13b778daa5c6
3
+ size 1444524
SoccerTwos/events.out.tfevents.1689930706.MBP-di-Giovanni.homenet.telecomitalia.it.8723.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d048b1053efe34ed0664661eb03b21cee8984edf774b0cdb86c4b6430b52754
3
+ size 1241
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SoccerTwos": {"trainer_type": "poca", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "constant", "beta_schedule": "constant", "epsilon_schedule": "constant"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 50000000, "time_horizon": 1000, "summary_freq": 10000, "threaded": false, "self_play": {"save_steps": 50000, "team_change": 200000, "swap_steps": 2000, "window": 10, "play_against_latest_model_ratio": 0.5, "initial_elo": 1200.0}, "behavioral_cloning": null}}, "env_settings": {"env_path": "training-envs-executables/linux/SoccerTwos/SoccerTwos", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "=SoccerTwos training", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SoccerTwos": {"trainer_type": "poca", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "constant", "beta_schedule": "constant", "epsilon_schedule": "constant"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 50000000, "time_horizon": 1000, "summary_freq": 10000, "threaded": false, "self_play": {"save_steps": 50000, "team_change": 200000, "swap_steps": 2000, "window": 10, "play_against_latest_model_ratio": 0.5, "initial_elo": 1200.0}, "behavioral_cloning": null}}, "env_settings": {"env_path": "training-envs-executables/linux/SoccerTwos/SoccerTwos", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "=SoccerTwos training", "initialize_from": null, "load_model": false, "resume": true, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -66,14 +66,14 @@ engine_settings:
66
  time_scale: 20
67
  target_frame_rate: -1
68
  capture_frame_rate: 60
69
- no_graphics: false
70
  environment_parameters: null
71
  checkpoint_settings:
72
  run_id: =SoccerTwos training
73
  initialize_from: null
74
  load_model: false
75
- resume: false
76
- force: true
77
  train_model: false
78
  inference: false
79
  results_dir: results
 
66
  time_scale: 20
67
  target_frame_rate: -1
68
  capture_frame_rate: 60
69
+ no_graphics: true
70
  environment_parameters: null
71
  checkpoint_settings:
72
  run_id: =SoccerTwos training
73
  initialize_from: null
74
  load_model: false
75
+ resume: true
76
+ force: false
77
  train_model: false
78
  inference: false
79
  results_dir: results
run_logs/Player-0.log CHANGED
@@ -10,7 +10,7 @@ NullGfxDevice:
10
  Renderer: Null Device
11
  Vendor: Unity Technologies
12
  Begin MonoManager ReloadAssembly
13
- - Completed reload, in 0.084 seconds
14
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
15
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
16
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -35,503 +35,6 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
35
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
36
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
37
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
38
- UnloadTime: 0.325959 ms
39
  Registered Communicator in Agent.
40
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
- TOO LONG
42
- TOO LONG
43
- TOO LONG
44
- TOO LONG
45
- TOO LONG
46
- TOO LONG
47
- TOO LONG
48
- TOO LONG
49
- TOO LONG
50
- TOO LONG
51
- TOO LONG
52
- TOO LONG
53
- TOO LONG
54
- TOO LONG
55
- TOO LONG
56
- TOO LONG
57
- TOO LONG
58
- TOO LONG
59
- TOO LONG
60
- TOO LONG
61
- TOO LONG
62
- TOO LONG
63
- TOO LONG
64
- TOO LONG
65
- TOO LONG
66
- TOO LONG
67
- TOO LONG
68
- TOO LONG
69
- TOO LONG
70
- TOO LONG
71
- TOO LONG
72
- TOO LONG
73
- TOO LONG
74
- TOO LONG
75
- TOO LONG
76
- TOO LONG
77
- TOO LONG
78
- TOO LONG
79
- TOO LONG
80
- TOO LONG
81
- TOO LONG
82
- TOO LONG
83
- TOO LONG
84
- TOO LONG
85
- TOO LONG
86
- TOO LONG
87
- TOO LONG
88
- TOO LONG
89
- TOO LONG
90
- TOO LONG
91
- TOO LONG
92
- TOO LONG
93
- TOO LONG
94
- TOO LONG
95
- TOO LONG
96
- TOO LONG
97
- TOO LONG
98
- TOO LONG
99
- TOO LONG
100
- TOO LONG
101
- TOO LONG
102
- TOO LONG
103
- TOO LONG
104
- TOO LONG
105
- TOO LONG
106
- TOO LONG
107
- TOO LONG
108
- TOO LONG
109
- TOO LONG
110
- TOO LONG
111
- TOO LONG
112
- TOO LONG
113
- TOO LONG
114
- TOO LONG
115
- TOO LONG
116
- TOO LONG
117
- TOO LONG
118
- TOO LONG
119
- TOO LONG
120
- TOO LONG
121
- TOO LONG
122
- TOO LONG
123
- TOO LONG
124
- TOO LONG
125
- TOO LONG
126
- TOO LONG
127
- TOO LONG
128
- TOO LONG
129
- TOO LONG
130
- TOO LONG
131
- TOO LONG
132
- TOO LONG
133
- TOO LONG
134
- TOO LONG
135
- TOO LONG
136
- TOO LONG
137
- TOO LONG
138
- TOO LONG
139
- TOO LONG
140
- TOO LONG
141
- TOO LONG
142
- TOO LONG
143
- TOO LONG
144
- TOO LONG
145
- TOO LONG
146
- TOO LONG
147
- TOO LONG
148
- TOO LONG
149
- TOO LONG
150
- TOO LONG
151
- TOO LONG
152
- TOO LONG
153
- TOO LONG
154
- TOO LONG
155
- TOO LONG
156
- TOO LONG
157
- TOO LONG
158
- TOO LONG
159
- TOO LONG
160
- TOO LONG
161
- TOO LONG
162
- TOO LONG
163
- TOO LONG
164
- TOO LONG
165
- TOO LONG
166
- TOO LONG
167
- TOO LONG
168
- TOO LONG
169
- TOO LONG
170
- TOO LONG
171
- TOO LONG
172
- TOO LONG
173
- TOO LONG
174
- TOO LONG
175
- TOO LONG
176
- TOO LONG
177
- TOO LONG
178
- TOO LONG
179
- TOO LONG
180
- TOO LONG
181
- TOO LONG
182
- TOO LONG
183
- TOO LONG
184
- TOO LONG
185
- TOO LONG
186
- TOO LONG
187
- TOO LONG
188
- TOO LONG
189
- TOO LONG
190
- TOO LONG
191
- TOO LONG
192
- TOO LONG
193
- TOO LONG
194
- TOO LONG
195
- TOO LONG
196
- TOO LONG
197
- TOO LONG
198
- TOO LONG
199
- TOO LONG
200
- TOO LONG
201
- TOO LONG
202
- TOO LONG
203
- TOO LONG
204
- TOO LONG
205
- TOO LONG
206
- TOO LONG
207
- TOO LONG
208
- TOO LONG
209
- TOO LONG
210
- TOO LONG
211
- TOO LONG
212
- TOO LONG
213
- TOO LONG
214
- TOO LONG
215
- TOO LONG
216
- TOO LONG
217
- TOO LONG
218
- TOO LONG
219
- TOO LONG
220
- TOO LONG
221
- TOO LONG
222
- TOO LONG
223
- TOO LONG
224
- TOO LONG
225
- TOO LONG
226
- TOO LONG
227
- TOO LONG
228
- TOO LONG
229
- TOO LONG
230
- TOO LONG
231
- TOO LONG
232
- TOO LONG
233
- TOO LONG
234
- TOO LONG
235
- TOO LONG
236
- TOO LONG
237
- TOO LONG
238
- TOO LONG
239
- TOO LONG
240
- TOO LONG
241
- TOO LONG
242
- TOO LONG
243
- TOO LONG
244
- TOO LONG
245
- TOO LONG
246
- TOO LONG
247
- TOO LONG
248
- TOO LONG
249
- TOO LONG
250
- TOO LONG
251
- TOO LONG
252
- TOO LONG
253
- TOO LONG
254
- TOO LONG
255
- TOO LONG
256
- TOO LONG
257
- TOO LONG
258
- TOO LONG
259
- TOO LONG
260
- TOO LONG
261
- TOO LONG
262
- TOO LONG
263
- TOO LONG
264
- TOO LONG
265
- TOO LONG
266
- TOO LONG
267
- TOO LONG
268
- TOO LONG
269
- TOO LONG
270
- TOO LONG
271
- TOO LONG
272
- TOO LONG
273
- TOO LONG
274
- TOO LONG
275
- TOO LONG
276
- TOO LONG
277
- TOO LONG
278
- TOO LONG
279
- TOO LONG
280
- TOO LONG
281
- TOO LONG
282
- TOO LONG
283
- TOO LONG
284
- TOO LONG
285
- TOO LONG
286
- TOO LONG
287
- TOO LONG
288
- TOO LONG
289
- TOO LONG
290
- TOO LONG
291
- TOO LONG
292
- TOO LONG
293
- TOO LONG
294
- TOO LONG
295
- TOO LONG
296
- TOO LONG
297
- TOO LONG
298
- TOO LONG
299
- TOO LONG
300
- TOO LONG
301
- TOO LONG
302
- TOO LONG
303
- TOO LONG
304
- TOO LONG
305
- TOO LONG
306
- TOO LONG
307
- TOO LONG
308
- TOO LONG
309
- TOO LONG
310
- TOO LONG
311
- TOO LONG
312
- TOO LONG
313
- TOO LONG
314
- TOO LONG
315
- TOO LONG
316
- TOO LONG
317
- TOO LONG
318
- TOO LONG
319
- TOO LONG
320
- TOO LONG
321
- TOO LONG
322
- TOO LONG
323
- TOO LONG
324
- TOO LONG
325
- TOO LONG
326
- TOO LONG
327
- TOO LONG
328
- TOO LONG
329
- TOO LONG
330
- TOO LONG
331
- TOO LONG
332
- TOO LONG
333
- TOO LONG
334
- TOO LONG
335
- TOO LONG
336
- TOO LONG
337
- TOO LONG
338
- TOO LONG
339
- TOO LONG
340
- TOO LONG
341
- TOO LONG
342
- TOO LONG
343
- TOO LONG
344
- TOO LONG
345
- TOO LONG
346
- TOO LONG
347
- TOO LONG
348
- TOO LONG
349
- TOO LONG
350
- TOO LONG
351
- TOO LONG
352
- TOO LONG
353
- TOO LONG
354
- TOO LONG
355
- TOO LONG
356
- TOO LONG
357
- TOO LONG
358
- TOO LONG
359
- TOO LONG
360
- TOO LONG
361
- TOO LONG
362
- TOO LONG
363
- TOO LONG
364
- TOO LONG
365
- TOO LONG
366
- TOO LONG
367
- TOO LONG
368
- TOO LONG
369
- TOO LONG
370
- TOO LONG
371
- TOO LONG
372
- TOO LONG
373
- TOO LONG
374
- TOO LONG
375
- TOO LONG
376
- TOO LONG
377
- TOO LONG
378
- TOO LONG
379
- TOO LONG
380
- TOO LONG
381
- TOO LONG
382
- TOO LONG
383
- TOO LONG
384
- TOO LONG
385
- TOO LONG
386
- TOO LONG
387
- TOO LONG
388
- TOO LONG
389
- TOO LONG
390
- TOO LONG
391
- TOO LONG
392
- TOO LONG
393
- TOO LONG
394
- TOO LONG
395
- TOO LONG
396
- TOO LONG
397
- TOO LONG
398
- TOO LONG
399
- TOO LONG
400
- TOO LONG
401
- TOO LONG
402
- TOO LONG
403
- TOO LONG
404
- TOO LONG
405
- TOO LONG
406
- TOO LONG
407
- TOO LONG
408
- TOO LONG
409
- TOO LONG
410
- TOO LONG
411
- TOO LONG
412
- TOO LONG
413
- TOO LONG
414
- TOO LONG
415
- TOO LONG
416
- TOO LONG
417
- TOO LONG
418
- TOO LONG
419
- TOO LONG
420
- TOO LONG
421
- TOO LONG
422
- TOO LONG
423
- TOO LONG
424
- TOO LONG
425
- TOO LONG
426
- TOO LONG
427
- TOO LONG
428
- TOO LONG
429
- TOO LONG
430
- TOO LONG
431
- TOO LONG
432
- TOO LONG
433
- TOO LONG
434
- TOO LONG
435
- TOO LONG
436
- TOO LONG
437
- TOO LONG
438
- TOO LONG
439
- TOO LONG
440
- TOO LONG
441
- TOO LONG
442
- TOO LONG
443
- TOO LONG
444
- TOO LONG
445
- TOO LONG
446
- TOO LONG
447
- TOO LONG
448
- TOO LONG
449
- TOO LONG
450
- TOO LONG
451
- TOO LONG
452
- TOO LONG
453
- TOO LONG
454
- TOO LONG
455
- TOO LONG
456
- TOO LONG
457
- TOO LONG
458
- TOO LONG
459
- TOO LONG
460
- TOO LONG
461
- TOO LONG
462
- TOO LONG
463
- TOO LONG
464
- TOO LONG
465
- TOO LONG
466
- TOO LONG
467
- TOO LONG
468
- TOO LONG
469
- TOO LONG
470
- TOO LONG
471
- TOO LONG
472
- TOO LONG
473
- TOO LONG
474
- TOO LONG
475
- TOO LONG
476
- TOO LONG
477
- TOO LONG
478
- TOO LONG
479
- TOO LONG
480
- TOO LONG
481
- TOO LONG
482
- TOO LONG
483
- TOO LONG
484
- TOO LONG
485
- TOO LONG
486
- TOO LONG
487
- TOO LONG
488
- TOO LONG
489
- TOO LONG
490
- TOO LONG
491
- TOO LONG
492
- TOO LONG
493
- TOO LONG
494
- TOO LONG
495
- TOO LONG
496
- TOO LONG
497
- TOO LONG
498
- TOO LONG
499
- TOO LONG
500
- TOO LONG
501
- TOO LONG
502
- TOO LONG
503
- TOO LONG
504
- TOO LONG
505
- TOO LONG
506
- TOO LONG
507
- TOO LONG
508
- TOO LONG
509
- TOO LONG
510
- TOO LONG
511
- TOO LONG
512
- TOO LONG
513
- TOO LONG
514
- TOO LONG
515
- TOO LONG
516
- TOO LONG
517
- TOO LONG
518
- TOO LONG
519
- TOO LONG
520
- TOO LONG
521
- TOO LONG
522
- TOO LONG
523
- TOO LONG
524
- TOO LONG
525
- TOO LONG
526
- TOO LONG
527
- TOO LONG
528
- TOO LONG
529
- TOO LONG
530
- TOO LONG
531
- TOO LONG
532
- TOO LONG
533
- TOO LONG
534
- TOO LONG
535
- TOO LONG
536
- TOO LONG
537
- TOO LONG
 
10
  Renderer: Null Device
11
  Vendor: Unity Technologies
12
  Begin MonoManager ReloadAssembly
13
+ - Completed reload, in 0.068 seconds
14
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
15
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
16
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
35
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
36
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
37
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
38
+ UnloadTime: 0.256916 ms
39
  Registered Communicator in Agent.
40
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_logs/timers.json CHANGED
@@ -2,305 +2,305 @@
2
  "name": "root",
3
  "gauges": {
4
  "SoccerTwos.Policy.Entropy.mean": {
5
- "value": 3.2730460166931152,
6
- "min": 3.2722370624542236,
7
- "max": 3.295734167098999,
8
- "count": 13
9
  },
10
  "SoccerTwos.Policy.Entropy.sum": {
11
- "value": 72373.59375,
12
- "min": 19247.623046875,
13
- "max": 105463.4765625,
14
- "count": 13
15
  },
16
  "SoccerTwos.Environment.EpisodeLength.mean": {
17
- "value": 830.0,
18
- "min": 727.5,
19
  "max": 999.0,
20
- "count": 13
21
  },
22
  "SoccerTwos.Environment.EpisodeLength.sum": {
23
- "value": 19920.0,
24
- "min": 16748.0,
25
- "max": 26100.0,
26
- "count": 13
27
  },
28
  "SoccerTwos.Self-play.ELO.mean": {
29
- "value": 1204.286574055704,
30
- "min": 1199.9602811060863,
31
- "max": 1204.286574055704,
32
- "count": 12
33
  },
34
  "SoccerTwos.Self-play.ELO.sum": {
35
- "value": 4817.146296222816,
36
- "min": 2402.3965187876356,
37
- "max": 9599.68224884869,
38
- "count": 12
 
 
 
 
 
 
 
 
 
 
 
 
39
  },
40
  "SoccerTwos.Step.mean": {
41
- "value": 129820.0,
42
- "min": 9066.0,
43
- "max": 129820.0,
44
- "count": 13
45
  },
46
  "SoccerTwos.Step.sum": {
47
- "value": 129820.0,
48
- "min": 9066.0,
49
- "max": 129820.0,
50
- "count": 13
51
  },
52
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
53
- "value": -0.0008173122187145054,
54
- "min": -0.00990950409322977,
55
- "max": 0.0052213906310498714,
56
- "count": 13
57
  },
58
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
59
- "value": -0.009807746857404709,
60
- "min": -0.12882354855537415,
61
- "max": 0.06265668570995331,
62
- "count": 13
63
  },
64
  "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
65
- "value": -0.0019189155427739024,
66
- "min": -0.00903354398906231,
67
- "max": 0.005207187030464411,
68
- "count": 13
69
  },
70
  "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
71
- "value": -0.02302698604762554,
72
- "min": -0.11743606626987457,
73
- "max": 0.06248624250292778,
74
- "count": 13
75
  },
76
  "SoccerTwos.Environment.CumulativeReward.mean": {
77
  "value": 0.0,
78
  "min": 0.0,
79
  "max": 0.0,
80
- "count": 13
81
  },
82
  "SoccerTwos.Environment.CumulativeReward.sum": {
83
  "value": 0.0,
84
  "min": 0.0,
85
  "max": 0.0,
86
- "count": 13
87
  },
88
  "SoccerTwos.Policy.ExtrinsicReward.mean": {
89
- "value": -0.061700006326039634,
90
- "min": -0.3194153859065129,
91
- "max": 0.20903334021568298,
92
- "count": 13
93
  },
94
  "SoccerTwos.Policy.ExtrinsicReward.sum": {
95
- "value": -0.7404000759124756,
96
- "min": -4.152400016784668,
97
- "max": 2.508400082588196,
98
- "count": 13
99
  },
100
  "SoccerTwos.Environment.GroupCumulativeReward.mean": {
101
- "value": -0.061700006326039634,
102
- "min": -0.3194153859065129,
103
- "max": 0.20903334021568298,
104
- "count": 13
105
  },
106
  "SoccerTwos.Environment.GroupCumulativeReward.sum": {
107
- "value": -0.7404000759124756,
108
- "min": -4.152400016784668,
109
- "max": 2.508400082588196,
110
- "count": 13
111
- },
112
- "SoccerTwos.IsTraining.mean": {
113
- "value": 1.0,
114
- "min": 1.0,
115
- "max": 1.0,
116
- "count": 13
117
- },
118
- "SoccerTwos.IsTraining.sum": {
119
- "value": 1.0,
120
- "min": 1.0,
121
- "max": 1.0,
122
- "count": 13
123
  },
124
  "SoccerTwos.Losses.PolicyLoss.mean": {
125
- "value": 0.012841316836420447,
126
- "min": 0.012841316836420447,
127
- "max": 0.017041035451984498,
128
- "count": 6
129
  },
130
  "SoccerTwos.Losses.PolicyLoss.sum": {
131
- "value": 0.012841316836420447,
132
- "min": 0.012841316836420447,
133
- "max": 0.017041035451984498,
134
- "count": 6
135
  },
136
  "SoccerTwos.Losses.ValueLoss.mean": {
137
- "value": 0.0021382930882585544,
138
- "min": 0.0016193612692101548,
139
- "max": 0.003799449202294151,
140
- "count": 6
141
  },
142
  "SoccerTwos.Losses.ValueLoss.sum": {
143
- "value": 0.0021382930882585544,
144
- "min": 0.0016193612692101548,
145
- "max": 0.003799449202294151,
146
- "count": 6
147
  },
148
  "SoccerTwos.Losses.BaselineLoss.mean": {
149
- "value": 0.0021427319734357297,
150
- "min": 0.0016194361300828557,
151
- "max": 0.003858683972309033,
152
- "count": 6
153
  },
154
  "SoccerTwos.Losses.BaselineLoss.sum": {
155
- "value": 0.0021427319734357297,
156
- "min": 0.0016194361300828557,
157
- "max": 0.003858683972309033,
158
- "count": 6
159
  },
160
  "SoccerTwos.Policy.LearningRate.mean": {
161
  "value": 0.0003,
162
  "min": 0.0003,
163
  "max": 0.0003,
164
- "count": 6
165
  },
166
  "SoccerTwos.Policy.LearningRate.sum": {
167
  "value": 0.0003,
168
  "min": 0.0003,
169
  "max": 0.0003,
170
- "count": 6
171
  },
172
  "SoccerTwos.Policy.Epsilon.mean": {
173
  "value": 0.20000000000000007,
174
  "min": 0.20000000000000007,
175
  "max": 0.20000000000000007,
176
- "count": 6
177
  },
178
  "SoccerTwos.Policy.Epsilon.sum": {
179
  "value": 0.20000000000000007,
180
  "min": 0.20000000000000007,
181
  "max": 0.20000000000000007,
182
- "count": 6
183
  },
184
  "SoccerTwos.Policy.Beta.mean": {
185
  "value": 0.005000000000000001,
186
  "min": 0.005000000000000001,
187
  "max": 0.005000000000000001,
188
- "count": 6
189
  },
190
  "SoccerTwos.Policy.Beta.sum": {
191
  "value": 0.005000000000000001,
192
  "min": 0.005000000000000001,
193
  "max": 0.005000000000000001,
194
- "count": 6
195
  }
196
  },
197
  "metadata": {
198
  "timer_format_version": "0.1.0",
199
- "start_time_seconds": "1689870004",
200
  "python_version": "3.9.16 | packaged by conda-forge | (main, Feb 1 2023, 21:38:11) \n[Clang 14.0.6 ]",
201
- "command_line_arguments": "/Users/giovannidispoto/miniforge3/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env training-envs-executables/linux/SoccerTwos/SoccerTwos --run-id =SoccerTwos training --force",
202
  "mlagents_version": "0.31.0.dev0",
203
  "mlagents_envs_version": "0.31.0.dev0",
204
  "communication_protocol_version": "1.5.0",
205
  "pytorch_version": "1.11.0",
206
  "numpy_version": "1.21.2",
207
- "end_time_seconds": "1689870497"
208
  },
209
- "total": 493.635118291,
210
  "count": 1,
211
- "self": 0.005367124000088097,
212
  "children": {
213
  "run_training.setup": {
214
- "total": 0.012915374999999951,
215
  "count": 1,
216
- "self": 0.012915374999999951
217
  },
218
  "TrainerController.start_learning": {
219
- "total": 493.61683579199996,
220
  "count": 1,
221
- "self": 0.09755877800296275,
222
  "children": {
223
  "TrainerController._reset_env": {
224
- "total": 3.325448333,
225
- "count": 1,
226
- "self": 3.325448333
227
  },
228
  "TrainerController.advance": {
229
- "total": 489.970734097997,
230
- "count": 8888,
231
- "self": 0.0966509849946533,
232
  "children": {
233
  "env_step": {
234
- "total": 396.18781395499997,
235
- "count": 8888,
236
- "self": 380.7007259920022,
237
  "children": {
238
  "SubprocessEnvManager._take_step": {
239
- "total": 15.422013874997418,
240
- "count": 8888,
241
- "self": 0.5089644479970659,
242
  "children": {
243
  "TorchPolicy.evaluate": {
244
- "total": 14.913049427000352,
245
- "count": 17648,
246
- "self": 14.913049427000352
247
  }
248
  }
249
  },
250
  "workers": {
251
- "total": 0.06507408800034975,
252
- "count": 8887,
253
  "self": 0.0,
254
  "children": {
255
  "worker_root": {
256
- "total": 486.1569761020012,
257
- "count": 8887,
258
  "is_parallel": true,
259
- "self": 122.89936898400254,
260
  "children": {
261
  "steps_from_proto": {
262
- "total": 0.0015613330000001646,
263
- "count": 2,
264
  "is_parallel": true,
265
- "self": 0.000280333000000077,
266
  "children": {
267
  "_process_rank_one_or_two_observation": {
268
- "total": 0.0012810000000000876,
269
- "count": 8,
270
  "is_parallel": true,
271
- "self": 0.0012810000000000876
272
  }
273
  }
274
  },
275
  "UnityEnvironment.step": {
276
- "total": 363.25604578499866,
277
- "count": 8887,
278
  "is_parallel": true,
279
- "self": 0.9835880289940064,
280
  "children": {
281
  "UnityEnvironment._generate_step_input": {
282
- "total": 5.796755179999298,
283
- "count": 8887,
284
  "is_parallel": true,
285
- "self": 5.796755179999298
286
  },
287
  "communicator.exchange": {
288
- "total": 345.07033209000235,
289
- "count": 8887,
290
  "is_parallel": true,
291
- "self": 345.07033209000235
292
  },
293
  "steps_from_proto": {
294
- "total": 11.405370486002987,
295
- "count": 17774,
296
  "is_parallel": true,
297
- "self": 1.5637338320070846,
298
  "children": {
299
  "_process_rank_one_or_two_observation": {
300
- "total": 9.841636653995902,
301
- "count": 71096,
302
  "is_parallel": true,
303
- "self": 9.841636653995902
304
  }
305
  }
306
  }
@@ -313,24 +313,31 @@
313
  }
314
  },
315
  "trainer_advance": {
316
- "total": 93.68626915800243,
317
- "count": 8887,
318
- "self": 0.7808035860016105,
319
  "children": {
320
  "process_trajectory": {
321
- "total": 13.667919948000808,
322
- "count": 8887,
323
- "self": 13.667919948000808
 
 
 
 
 
 
 
324
  },
325
  "_update_policy": {
326
- "total": 79.237545624,
327
- "count": 6,
328
- "self": 10.890855251999994,
329
  "children": {
330
  "TorchPOCAOptimizer.update": {
331
- "total": 68.34669037200001,
332
- "count": 180,
333
- "self": 68.34669037200001
334
  }
335
  }
336
  }
@@ -339,19 +346,19 @@
339
  }
340
  },
341
  "trainer_threads": {
342
- "total": 2.7499999987412593e-05,
343
  "count": 1,
344
- "self": 2.7499999987412593e-05
345
  },
346
  "TrainerController._save_models": {
347
- "total": 0.22306708299998945,
348
  "count": 1,
349
- "self": 0.0021550000000161162,
350
  "children": {
351
  "RLTrainer._checkpoint": {
352
- "total": 0.22091208299997334,
353
  "count": 1,
354
- "self": 0.22091208299997334
355
  }
356
  }
357
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SoccerTwos.Policy.Entropy.mean": {
5
+ "value": 3.0429799556732178,
6
+ "min": 3.0280234813690186,
7
+ "max": 3.191495418548584,
8
+ "count": 161
9
  },
10
  "SoccerTwos.Policy.Entropy.sum": {
11
+ "value": 55893.45703125,
12
+ "min": 9785.7265625,
13
+ "max": 110518.34375,
14
+ "count": 161
15
  },
16
  "SoccerTwos.Environment.EpisodeLength.mean": {
17
+ "value": 573.5714285714286,
18
+ "min": 95.0,
19
  "max": 999.0,
20
+ "count": 161
21
  },
22
  "SoccerTwos.Environment.EpisodeLength.sum": {
23
+ "value": 16060.0,
24
+ "min": 380.0,
25
+ "max": 23704.0,
26
+ "count": 161
27
  },
28
  "SoccerTwos.Self-play.ELO.mean": {
29
+ "value": 1253.5616176330443,
30
+ "min": 1198.1748449658994,
31
+ "max": 1255.1449382701196,
32
+ "count": 153
33
  },
34
  "SoccerTwos.Self-play.ELO.sum": {
35
+ "value": 12535.616176330443,
36
+ "min": 2396.349689931799,
37
+ "max": 45056.64592682517,
38
+ "count": 153
39
+ },
40
+ "SoccerTwos.IsTraining.mean": {
41
+ "value": 1.0,
42
+ "min": 1.0,
43
+ "max": 1.0,
44
+ "count": 161
45
+ },
46
+ "SoccerTwos.IsTraining.sum": {
47
+ "value": 1.0,
48
+ "min": 1.0,
49
+ "max": 1.0,
50
+ "count": 161
51
  },
52
  "SoccerTwos.Step.mean": {
53
+ "value": 2599727.0,
54
+ "min": 1009454.0,
55
+ "max": 2599727.0,
56
+ "count": 160
57
  },
58
  "SoccerTwos.Step.sum": {
59
+ "value": 2599727.0,
60
+ "min": 1009454.0,
61
+ "max": 2599727.0,
62
+ "count": 160
63
  },
64
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
65
+ "value": -0.015503649599850178,
66
+ "min": -0.04074995219707489,
67
+ "max": 0.03584587574005127,
68
+ "count": 160
69
  },
70
  "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
71
+ "value": -0.23255474865436554,
72
+ "min": -0.6843217611312866,
73
+ "max": 1.254599928855896,
74
+ "count": 160
75
  },
76
  "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
77
+ "value": -0.015258976258337498,
78
+ "min": -0.039837684482336044,
79
+ "max": 0.03866251930594444,
80
+ "count": 160
81
  },
82
  "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
83
+ "value": -0.22888463735580444,
84
+ "min": -0.6837544441223145,
85
+ "max": 1.353188157081604,
86
+ "count": 160
87
  },
88
  "SoccerTwos.Environment.CumulativeReward.mean": {
89
  "value": 0.0,
90
  "min": 0.0,
91
  "max": 0.0,
92
+ "count": 160
93
  },
94
  "SoccerTwos.Environment.CumulativeReward.sum": {
95
  "value": 0.0,
96
  "min": 0.0,
97
  "max": 0.0,
98
+ "count": 160
99
  },
100
  "SoccerTwos.Policy.ExtrinsicReward.mean": {
101
+ "value": -0.2710533301035563,
102
+ "min": -0.7474956512451172,
103
+ "max": 0.46924210222143875,
104
+ "count": 160
105
  },
106
  "SoccerTwos.Policy.ExtrinsicReward.sum": {
107
+ "value": -4.065799951553345,
108
+ "min": -17.192399978637695,
109
+ "max": 9.174400091171265,
110
+ "count": 160
111
  },
112
  "SoccerTwos.Environment.GroupCumulativeReward.mean": {
113
+ "value": -0.2710533301035563,
114
+ "min": -0.7474956512451172,
115
+ "max": 0.46924210222143875,
116
+ "count": 160
117
  },
118
  "SoccerTwos.Environment.GroupCumulativeReward.sum": {
119
+ "value": -4.065799951553345,
120
+ "min": -17.192399978637695,
121
+ "max": 9.174400091171265,
122
+ "count": 160
 
 
 
 
 
 
 
 
 
 
 
 
123
  },
124
  "SoccerTwos.Losses.PolicyLoss.mean": {
125
+ "value": 0.01552068288049971,
126
+ "min": 0.012524987536016852,
127
+ "max": 0.022274984690981606,
128
+ "count": 75
129
  },
130
  "SoccerTwos.Losses.PolicyLoss.sum": {
131
+ "value": 0.01552068288049971,
132
+ "min": 0.012524987536016852,
133
+ "max": 0.022274984690981606,
134
+ "count": 75
135
  },
136
  "SoccerTwos.Losses.ValueLoss.mean": {
137
+ "value": 0.012713205628097057,
138
+ "min": 0.0005046518689293104,
139
+ "max": 0.015418903343379498,
140
+ "count": 75
141
  },
142
  "SoccerTwos.Losses.ValueLoss.sum": {
143
+ "value": 0.012713205628097057,
144
+ "min": 0.0005046518689293104,
145
+ "max": 0.015418903343379498,
146
+ "count": 75
147
  },
148
  "SoccerTwos.Losses.BaselineLoss.mean": {
149
+ "value": 0.01291948426514864,
150
+ "min": 0.000504054954944877,
151
+ "max": 0.01562350361297528,
152
+ "count": 75
153
  },
154
  "SoccerTwos.Losses.BaselineLoss.sum": {
155
+ "value": 0.01291948426514864,
156
+ "min": 0.000504054954944877,
157
+ "max": 0.01562350361297528,
158
+ "count": 75
159
  },
160
  "SoccerTwos.Policy.LearningRate.mean": {
161
  "value": 0.0003,
162
  "min": 0.0003,
163
  "max": 0.0003,
164
+ "count": 75
165
  },
166
  "SoccerTwos.Policy.LearningRate.sum": {
167
  "value": 0.0003,
168
  "min": 0.0003,
169
  "max": 0.0003,
170
+ "count": 75
171
  },
172
  "SoccerTwos.Policy.Epsilon.mean": {
173
  "value": 0.20000000000000007,
174
  "min": 0.20000000000000007,
175
  "max": 0.20000000000000007,
176
+ "count": 75
177
  },
178
  "SoccerTwos.Policy.Epsilon.sum": {
179
  "value": 0.20000000000000007,
180
  "min": 0.20000000000000007,
181
  "max": 0.20000000000000007,
182
+ "count": 75
183
  },
184
  "SoccerTwos.Policy.Beta.mean": {
185
  "value": 0.005000000000000001,
186
  "min": 0.005000000000000001,
187
  "max": 0.005000000000000001,
188
+ "count": 75
189
  },
190
  "SoccerTwos.Policy.Beta.sum": {
191
  "value": 0.005000000000000001,
192
  "min": 0.005000000000000001,
193
  "max": 0.005000000000000001,
194
+ "count": 75
195
  }
196
  },
197
  "metadata": {
198
  "timer_format_version": "0.1.0",
199
+ "start_time_seconds": "1689924786",
200
  "python_version": "3.9.16 | packaged by conda-forge | (main, Feb 1 2023, 21:38:11) \n[Clang 14.0.6 ]",
201
+ "command_line_arguments": "/Users/giovannidispoto/miniforge3/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env training-envs-executables/linux/SoccerTwos/SoccerTwos --run-id =SoccerTwos training --no-graphics --resume",
202
  "mlagents_version": "0.31.0.dev0",
203
  "mlagents_envs_version": "0.31.0.dev0",
204
  "communication_protocol_version": "1.5.0",
205
  "pytorch_version": "1.11.0",
206
  "numpy_version": "1.21.2",
207
+ "end_time_seconds": "1689930698"
208
  },
209
+ "total": 5911.306915667001,
210
  "count": 1,
211
+ "self": 0.20646262500122248,
212
  "children": {
213
  "run_training.setup": {
214
+ "total": 0.056267874999999856,
215
  "count": 1,
216
+ "self": 0.056267874999999856
217
  },
218
  "TrainerController.start_learning": {
219
+ "total": 5911.044185166999,
220
  "count": 1,
221
+ "self": 1.1270955429890819,
222
  "children": {
223
  "TrainerController._reset_env": {
224
+ "total": 2.8766715400011753,
225
+ "count": 10,
226
+ "self": 2.8766715400011753
227
  },
228
  "TrainerController.advance": {
229
+ "total": 5906.894979584009,
230
+ "count": 104383,
231
+ "self": 1.0448963620874565,
232
  "children": {
233
  "env_step": {
234
+ "total": 4601.6896993350165,
235
+ "count": 104383,
236
+ "self": 4409.8189295159245,
237
  "children": {
238
  "SubprocessEnvManager._take_step": {
239
+ "total": 191.10761183097588,
240
+ "count": 104383,
241
+ "self": 5.378499134121967,
242
  "children": {
243
  "TorchPolicy.evaluate": {
244
+ "total": 185.72911269685392,
245
+ "count": 206502,
246
+ "self": 185.72911269685392
247
  }
248
  }
249
  },
250
  "workers": {
251
+ "total": 0.7631579881156241,
252
+ "count": 104382,
253
  "self": 0.0,
254
  "children": {
255
  "worker_root": {
256
+ "total": 5906.829998433097,
257
+ "count": 104382,
258
  "is_parallel": true,
259
+ "self": 1679.6308405951095,
260
  "children": {
261
  "steps_from_proto": {
262
+ "total": 0.013667627000538918,
263
+ "count": 20,
264
  "is_parallel": true,
265
+ "self": 0.0020639220016827053,
266
  "children": {
267
  "_process_rank_one_or_two_observation": {
268
+ "total": 0.011603704998856212,
269
+ "count": 80,
270
  "is_parallel": true,
271
+ "self": 0.011603704998856212
272
  }
273
  }
274
  },
275
  "UnityEnvironment.step": {
276
+ "total": 4227.185490210987,
277
+ "count": 104382,
278
  "is_parallel": true,
279
+ "self": 11.35144842086902,
280
  "children": {
281
  "UnityEnvironment._generate_step_input": {
282
+ "total": 66.85198158409172,
283
+ "count": 104382,
284
  "is_parallel": true,
285
+ "self": 66.85198158409172
286
  },
287
  "communicator.exchange": {
288
+ "total": 4017.812950275919,
289
+ "count": 104382,
290
  "is_parallel": true,
291
+ "self": 4017.812950275919
292
  },
293
  "steps_from_proto": {
294
+ "total": 131.16910993010734,
295
+ "count": 208764,
296
  "is_parallel": true,
297
+ "self": 17.862625829338057,
298
  "children": {
299
  "_process_rank_one_or_two_observation": {
300
+ "total": 113.30648410076928,
301
+ "count": 835056,
302
  "is_parallel": true,
303
+ "self": 113.30648410076928
304
  }
305
  }
306
  }
 
313
  }
314
  },
315
  "trainer_advance": {
316
+ "total": 1304.1603838869046,
317
+ "count": 104382,
318
+ "self": 9.110193707981125,
319
  "children": {
320
  "process_trajectory": {
321
+ "total": 190.7648718849274,
322
+ "count": 104382,
323
+ "self": 190.26279075992727,
324
+ "children": {
325
+ "RLTrainer._checkpoint": {
326
+ "total": 0.5020811250001316,
327
+ "count": 4,
328
+ "self": 0.5020811250001316
329
+ }
330
+ }
331
  },
332
  "_update_policy": {
333
+ "total": 1104.2853182939962,
334
+ "count": 75,
335
+ "self": 136.70639390800068,
336
  "children": {
337
  "TorchPOCAOptimizer.update": {
338
+ "total": 967.5789243859955,
339
+ "count": 2250,
340
+ "self": 967.5789243859955
341
  }
342
  }
343
  }
 
346
  }
347
  },
348
  "trainer_threads": {
349
+ "total": 5.000001692678779e-07,
350
  "count": 1,
351
+ "self": 5.000001692678779e-07
352
  },
353
  "TrainerController._save_models": {
354
+ "total": 0.1454379999995581,
355
  "count": 1,
356
+ "self": 0.003915333999430004,
357
  "children": {
358
  "RLTrainer._checkpoint": {
359
+ "total": 0.14152266600012808,
360
  "count": 1,
361
+ "self": 0.14152266600012808
362
  }
363
  }
364
  }
run_logs/training_status.json CHANGED
@@ -1,24 +1,60 @@
1
  {
2
  "SoccerTwos": {
3
- "elo": 1204.896264587993,
4
  "checkpoints": [
5
  {
6
- "steps": 133820,
7
- "file_path": "results/=SoccerTwos training/SoccerTwos/SoccerTwos-133820.onnx",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "reward": 0.0,
9
- "creation_time": 1689870497.8246841,
10
  "auxillary_file_paths": [
11
- "results/=SoccerTwos training/SoccerTwos/SoccerTwos-133820.pt"
12
  ]
13
  }
14
  ],
15
  "final_checkpoint": {
16
- "steps": 133820,
17
  "file_path": "results/=SoccerTwos training/SoccerTwos.onnx",
18
  "reward": 0.0,
19
- "creation_time": 1689870497.8246841,
20
  "auxillary_file_paths": [
21
- "results/=SoccerTwos training/SoccerTwos/SoccerTwos-133820.pt"
22
  ]
23
  }
24
  },
 
1
  {
2
  "SoccerTwos": {
3
+ "elo": 1253.7885927395525,
4
  "checkpoints": [
5
  {
6
+ "steps": 999938,
7
+ "file_path": "results/=SoccerTwos training/SoccerTwos/SoccerTwos-999938.onnx",
8
+ "reward": null,
9
+ "creation_time": 1689924793.830441,
10
+ "auxillary_file_paths": [
11
+ "results/=SoccerTwos training/SoccerTwos/SoccerTwos-999938.pt"
12
+ ]
13
+ },
14
+ {
15
+ "steps": 1499744,
16
+ "file_path": "results/=SoccerTwos training/SoccerTwos/SoccerTwos-1499744.onnx",
17
+ "reward": 0.0,
18
+ "creation_time": 1689926636.394656,
19
+ "auxillary_file_paths": [
20
+ "results/=SoccerTwos training/SoccerTwos/SoccerTwos-1499744.pt"
21
+ ]
22
+ },
23
+ {
24
+ "steps": 1999130,
25
+ "file_path": "results/=SoccerTwos training/SoccerTwos/SoccerTwos-1999130.onnx",
26
+ "reward": 0.0,
27
+ "creation_time": 1689928436.199894,
28
+ "auxillary_file_paths": [
29
+ "results/=SoccerTwos training/SoccerTwos/SoccerTwos-1999130.pt"
30
+ ]
31
+ },
32
+ {
33
+ "steps": 2499468,
34
+ "file_path": "results/=SoccerTwos training/SoccerTwos/SoccerTwos-2499468.onnx",
35
+ "reward": 0.0,
36
+ "creation_time": 1689930295.224205,
37
+ "auxillary_file_paths": [
38
+ "results/=SoccerTwos training/SoccerTwos/SoccerTwos-2499468.pt"
39
+ ]
40
+ },
41
+ {
42
+ "steps": 2606308,
43
+ "file_path": "results/=SoccerTwos training/SoccerTwos/SoccerTwos-2606308.onnx",
44
  "reward": 0.0,
45
+ "creation_time": 1689930697.907537,
46
  "auxillary_file_paths": [
47
+ "results/=SoccerTwos training/SoccerTwos/SoccerTwos-2606308.pt"
48
  ]
49
  }
50
  ],
51
  "final_checkpoint": {
52
+ "steps": 2606308,
53
  "file_path": "results/=SoccerTwos training/SoccerTwos.onnx",
54
  "reward": 0.0,
55
+ "creation_time": 1689930697.907537,
56
  "auxillary_file_paths": [
57
+ "results/=SoccerTwos training/SoccerTwos/SoccerTwos-2606308.pt"
58
  ]
59
  }
60
  },