poca-SoccerTwos / run_logs /timers.json
AntiSquid's picture
First Push
6efd5c2
raw
history blame
20.2 kB
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.8680460453033447,
"min": 1.8161991834640503,
"max": 3.2957587242126465,
"count": 500
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 35507.8203125,
"min": 25353.796875,
"max": 133209.4375,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 54.0,
"min": 40.01652892561984,
"max": 999.0,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19872.0,
"min": 16412.0,
"max": 23464.0,
"count": 500
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1487.996427335391,
"min": 1198.3870190714101,
"max": 1499.079929772121,
"count": 487
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 273791.34262971196,
"min": 2398.1197929693635,
"max": 343898.34616915707,
"count": 487
},
"SoccerTwos.Step.mean": {
"value": 4999988.0,
"min": 9474.0,
"max": 4999988.0,
"count": 500
},
"SoccerTwos.Step.sum": {
"value": 4999988.0,
"min": 9474.0,
"max": 4999988.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.03525223210453987,
"min": -0.08694857358932495,
"max": 0.16366691887378693,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 6.45115852355957,
"min": -15.542802810668945,
"max": 29.951045989990234,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.034298527985811234,
"min": -0.08384038507938385,
"max": 0.15723483264446259,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 6.276630401611328,
"min": -15.197043418884277,
"max": 28.81496810913086,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.08900109275442655,
"min": -0.6129750004038215,
"max": 0.39745499938726425,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 16.28719997406006,
"min": -53.56880009174347,
"max": 49.409400165081024,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.08900109275442655,
"min": -0.6129750004038215,
"max": 0.39745499938726425,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 16.28719997406006,
"min": -53.56880009174347,
"max": 49.409400165081024,
"count": 500
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.017184486032541222,
"min": 0.010216441435234931,
"max": 0.023511711189833782,
"count": 240
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.017184486032541222,
"min": 0.010216441435234931,
"max": 0.023511711189833782,
"count": 240
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.10323547596732775,
"min": 3.487301008438711e-06,
"max": 0.11636010259389877,
"count": 240
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.10323547596732775,
"min": 3.487301008438711e-06,
"max": 0.11636010259389877,
"count": 240
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.10511516084273656,
"min": 4.465434430282282e-06,
"max": 0.12019507636626561,
"count": 240
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.10511516084273656,
"min": 4.465434430282282e-06,
"max": 0.12019507636626561,
"count": 240
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 240
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 240
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 240
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 240
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 240
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 240
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1678299642",
"python_version": "3.9.12 (main, Apr 5 2022, 06:56:58) \n[GCC 7.5.0]",
"command_line_arguments": "/home/autowin/anaconda3/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics --force",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1678307406"
},
"total": 7764.105665903,
"count": 1,
"self": 0.21977347299980465,
"children": {
"run_training.setup": {
"total": 0.026289096000255086,
"count": 1,
"self": 0.026289096000255086
},
"TrainerController.start_learning": {
"total": 7763.859603334,
"count": 1,
"self": 6.232340204809589,
"children": {
"TrainerController._reset_env": {
"total": 29.08605214200361,
"count": 25,
"self": 29.08605214200361
},
"TrainerController.advance": {
"total": 7728.278278909191,
"count": 340220,
"self": 5.822002416822215,
"children": {
"env_step": {
"total": 6027.354309442668,
"count": 340220,
"self": 4838.456784111771,
"children": {
"SubprocessEnvManager._take_step": {
"total": 1184.7400654937082,
"count": 340220,
"self": 29.90780162030933,
"children": {
"TorchPolicy.evaluate": {
"total": 1154.832263873399,
"count": 633256,
"self": 1154.832263873399
}
}
},
"workers": {
"total": 4.157459837188071,
"count": 340220,
"self": 0.0,
"children": {
"worker_root": {
"total": 7752.400014725616,
"count": 340220,
"is_parallel": true,
"self": 3687.2473595851297,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.006026917999406578,
"count": 2,
"is_parallel": true,
"self": 0.004680103000282543,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013468149991240352,
"count": 8,
"is_parallel": true,
"self": 0.0013468149991240352
}
}
},
"UnityEnvironment.step": {
"total": 0.024034647998632863,
"count": 1,
"is_parallel": true,
"self": 0.0007538039972132538,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.000784100000601029,
"count": 1,
"is_parallel": true,
"self": 0.000784100000601029
},
"communicator.exchange": {
"total": 0.02015413299886859,
"count": 1,
"is_parallel": true,
"self": 0.02015413299886859
},
"steps_from_proto": {
"total": 0.00234261100194999,
"count": 2,
"is_parallel": true,
"self": 0.0004599229978339281,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001882688004116062,
"count": 8,
"is_parallel": true,
"self": 0.001882688004116062
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 4065.1019752714856,
"count": 340219,
"is_parallel": true,
"self": 226.5033442609274,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 181.1389672658006,
"count": 340219,
"is_parallel": true,
"self": 181.1389672658006
},
"communicator.exchange": {
"total": 2948.8488457228304,
"count": 340219,
"is_parallel": true,
"self": 2948.8488457228304
},
"steps_from_proto": {
"total": 708.6108180219271,
"count": 680438,
"is_parallel": true,
"self": 132.42580274276042,
"children": {
"_process_rank_one_or_two_observation": {
"total": 576.1850152791667,
"count": 2721752,
"is_parallel": true,
"self": 576.1850152791667
}
}
}
}
},
"steps_from_proto": {
"total": 0.05067986900030519,
"count": 48,
"is_parallel": true,
"self": 0.009389518010721076,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.04129035098958411,
"count": 192,
"is_parallel": true,
"self": 0.04129035098958411
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 1695.1019670497008,
"count": 340220,
"self": 54.152541346084035,
"children": {
"process_trajectory": {
"total": 589.3418424656375,
"count": 340220,
"self": 586.7989732866372,
"children": {
"RLTrainer._checkpoint": {
"total": 2.542869179000263,
"count": 10,
"self": 2.542869179000263
}
}
},
"_update_policy": {
"total": 1051.6075832379793,
"count": 240,
"self": 687.9969801387997,
"children": {
"TorchPOCAOptimizer.update": {
"total": 363.61060309917957,
"count": 7200,
"self": 363.61060309917957
}
}
}
}
}
}
},
"trainer_threads": {
"total": 6.309965101536363e-07,
"count": 1,
"self": 6.309965101536363e-07
},
"TrainerController._save_models": {
"total": 0.2629314469995734,
"count": 1,
"self": 0.0015233880003506783,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2614080589992227,
"count": 1,
"self": 0.2614080589992227
}
}
}
}
}
}
}