ShreyasM's picture
First Push
454551a
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 2.12961745262146,
"min": 2.0128607749938965,
"max": 3.2957048416137695,
"count": 500
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 47294.546875,
"min": 25215.341796875,
"max": 114234.984375,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 85.3103448275862,
"min": 65.40789473684211,
"max": 999.0,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19792.0,
"min": 9760.0,
"max": 29316.0,
"count": 500
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1385.1089533898953,
"min": 1195.1336256730665,
"max": 1421.5866581636521,
"count": 455
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 160672.63859322786,
"min": 2392.120072047985,
"max": 214171.3717349403,
"count": 455
},
"SoccerTwos.Step.mean": {
"value": 4999996.0,
"min": 9398.0,
"max": 4999996.0,
"count": 500
},
"SoccerTwos.Step.sum": {
"value": 4999996.0,
"min": 9398.0,
"max": 4999996.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.01985156163573265,
"min": -0.16638648509979248,
"max": 0.19962865114212036,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 2.302781105041504,
"min": -13.310918807983398,
"max": 21.16063690185547,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.022537339478731155,
"min": -0.16885781288146973,
"max": 0.19671636819839478,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 2.6143314838409424,
"min": -13.508625030517578,
"max": 20.8519344329834,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.11248275843159906,
"min": -0.625,
"max": 0.4899577435473321,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 13.04799997806549,
"min": -35.64599967002869,
"max": 38.842000007629395,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.11248275843159906,
"min": -0.625,
"max": 0.4899577435473321,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 13.04799997806549,
"min": -35.64599967002869,
"max": 38.842000007629395,
"count": 500
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.034032461155826844,
"min": 0.01110359470185358,
"max": 0.03661396849978094,
"count": 239
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.034032461155826844,
"min": 0.01110359470185358,
"max": 0.03661396849978094,
"count": 239
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.05795758267243703,
"min": 9.818604667088948e-06,
"max": 4.2202854732672375,
"count": 239
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.05795758267243703,
"min": 9.818604667088948e-06,
"max": 4.2202854732672375,
"count": 239
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.058396623780330024,
"min": 6.856276786493254e-06,
"max": 3.7242351276800036,
"count": 239
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.058396623780330024,
"min": 6.856276786493254e-06,
"max": 3.7242351276800036,
"count": 239
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.003000000000000001,
"min": 0.0030000000000000005,
"max": 0.003000000000000001,
"count": 239
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.003000000000000001,
"min": 0.0030000000000000005,
"max": 0.003000000000000001,
"count": 239
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 239
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 239
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 239
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 239
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1698455272",
"python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
"command_line_arguments": "/home/acm/anaconda3/envs/ML-Agents/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos-AI --no-graphics --force",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu113",
"numpy_version": "1.21.2",
"end_time_seconds": "1698459197"
},
"total": 3925.1916756709106,
"count": 1,
"self": 0.16659858589991927,
"children": {
"run_training.setup": {
"total": 0.004986475920304656,
"count": 1,
"self": 0.004986475920304656
},
"TrainerController.start_learning": {
"total": 3925.0200906090904,
"count": 1,
"self": 2.2291876182425767,
"children": {
"TrainerController._reset_env": {
"total": 2.799344047671184,
"count": 25,
"self": 2.799344047671184
},
"TrainerController.advance": {
"total": 3919.90416892129,
"count": 330773,
"self": 2.2556251359637827,
"children": {
"env_step": {
"total": 3191.1467194603756,
"count": 330773,
"self": 2677.3377099183854,
"children": {
"SubprocessEnvManager._take_step": {
"total": 512.3794183228165,
"count": 330773,
"self": 17.684439467266202,
"children": {
"TorchPolicy.evaluate": {
"total": 494.6949788555503,
"count": 634300,
"self": 494.6949788555503
}
}
},
"workers": {
"total": 1.4295912191737443,
"count": 330773,
"self": 0.0,
"children": {
"worker_root": {
"total": 3915.8305289507844,
"count": 330773,
"is_parallel": true,
"self": 1587.0276870804373,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0012380797415971756,
"count": 2,
"is_parallel": true,
"self": 0.0002693717833608389,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0009687079582363367,
"count": 8,
"is_parallel": true,
"self": 0.0009687079582363367
}
}
},
"UnityEnvironment.step": {
"total": 0.012806799029931426,
"count": 1,
"is_parallel": true,
"self": 0.0002779781352728605,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00026643695309758186,
"count": 1,
"is_parallel": true,
"self": 0.00026643695309758186
},
"communicator.exchange": {
"total": 0.01133758807554841,
"count": 1,
"is_parallel": true,
"self": 0.01133758807554841
},
"steps_from_proto": {
"total": 0.0009247958660125732,
"count": 2,
"is_parallel": true,
"self": 0.00017542671412229538,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0007493691518902779,
"count": 8,
"is_parallel": true,
"self": 0.0007493691518902779
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 2328.770504897693,
"count": 330772,
"is_parallel": true,
"self": 132.92572601302527,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 90.27883003675379,
"count": 330772,
"is_parallel": true,
"self": 90.27883003675379
},
"communicator.exchange": {
"total": 1675.3324669057038,
"count": 330772,
"is_parallel": true,
"self": 1675.3324669057038
},
"steps_from_proto": {
"total": 430.23348194221035,
"count": 661544,
"is_parallel": true,
"self": 69.62733378470875,
"children": {
"_process_rank_one_or_two_observation": {
"total": 360.6061481575016,
"count": 2646176,
"is_parallel": true,
"self": 360.6061481575016
}
}
}
}
},
"steps_from_proto": {
"total": 0.032336972653865814,
"count": 48,
"is_parallel": true,
"self": 0.005629630060866475,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.02670734259299934,
"count": 192,
"is_parallel": true,
"self": 0.02670734259299934
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 726.5018243249506,
"count": 330773,
"self": 24.77505039377138,
"children": {
"process_trajectory": {
"total": 230.86068614642136,
"count": 330773,
"self": 229.9018766155932,
"children": {
"RLTrainer._checkpoint": {
"total": 0.9588095308281481,
"count": 10,
"self": 0.9588095308281481
}
}
},
"_update_policy": {
"total": 470.86608778475784,
"count": 239,
"self": 342.3568942944985,
"children": {
"TorchPOCAOptimizer.update": {
"total": 128.50919349025935,
"count": 7176,
"self": 128.50919349025935
}
}
}
}
}
}
},
"trainer_threads": {
"total": 4.409812390804291e-07,
"count": 1,
"self": 4.409812390804291e-07
},
"TrainerController._save_models": {
"total": 0.08738958090543747,
"count": 1,
"self": 0.0005776879843324423,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08681189292110503,
"count": 1,
"self": 0.08681189292110503
}
}
}
}
}
}
}