poca-SoccerTwos / run_logs /timers.json
danlindb's picture
First Push
2704a86
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 2.5264785289764404,
"min": 2.327571153640747,
"max": 3.295673131942749,
"count": 500
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 43253.3125,
"min": 24640.5546875,
"max": 132413.875,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 999.0,
"min": 414.92857142857144,
"max": 999.0,
"count": 500
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19980.0,
"min": 7992.0,
"max": 29880.0,
"count": 500
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1181.1800900094406,
"min": 1181.1645534042862,
"max": 1200.2986133514137,
"count": 188
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 2362.3601800188812,
"min": 2362.3291068085723,
"max": 16790.531359038996,
"count": 188
},
"SoccerTwos.Step.mean": {
"value": 4999270.0,
"min": 9958.0,
"max": 4999270.0,
"count": 500
},
"SoccerTwos.Step.sum": {
"value": 4999270.0,
"min": 9958.0,
"max": 4999270.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 2.0931589460815303e-05,
"min": -0.021131739020347595,
"max": 0.018906112760305405,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 0.00020931589824613184,
"min": -0.2585718631744385,
"max": 0.2431655079126358,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.0850846340181306e-05,
"min": -0.020071206614375114,
"max": 0.018901551142334938,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 0.00020850847067777067,
"min": -0.26092568039894104,
"max": 0.23107841610908508,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.0,
"min": -0.5714285714285714,
"max": 0.4275124967098236,
"count": 500
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 0.0,
"min": -8.0,
"max": 6.840199947357178,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.0,
"min": -0.5714285714285714,
"max": 0.4275124967098236,
"count": 500
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 0.0,
"min": -8.0,
"max": 6.840199947357178,
"count": 500
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 500
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.017995319445617497,
"min": 0.010117436539258052,
"max": 0.026128154496351877,
"count": 230
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.017995319445617497,
"min": 0.010117436539258052,
"max": 0.026128154496351877,
"count": 230
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 1.0822907311099547e-08,
"min": 1.1835150896348048e-09,
"max": 0.0059672161936759945,
"count": 230
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 1.0822907311099547e-08,
"min": 1.1835150896348048e-09,
"max": 0.0059672161936759945,
"count": 230
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 1.5197493204629117e-08,
"min": 1.4181593011031169e-09,
"max": 0.006036105689903101,
"count": 230
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 1.5197493204629117e-08,
"min": 1.4181593011031169e-09,
"max": 0.006036105689903101,
"count": 230
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 230
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 230
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 230
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 230
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 230
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 230
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1704546429",
"python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
"command_line_arguments": "/home/dlindbla/miniconda3/envs/HuggingFace-rl/bin/mlagents-learn config/poca/SoccerTwos.yaml --env=training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos-a1 --no-graphics --force",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.1.2+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1704558166"
},
"total": 11737.46440943,
"count": 1,
"self": 0.5942679859999771,
"children": {
"run_training.setup": {
"total": 0.018596790000174224,
"count": 1,
"self": 0.018596790000174224
},
"TrainerController.start_learning": {
"total": 11736.851544654,
"count": 1,
"self": 6.058674017456724,
"children": {
"TrainerController._reset_env": {
"total": 4.588709092999579,
"count": 25,
"self": 4.588709092999579
},
"TrainerController.advance": {
"total": 11725.882777993545,
"count": 325557,
"self": 7.259640768743338,
"children": {
"env_step": {
"total": 9806.669946348145,
"count": 325557,
"self": 6450.214249978521,
"children": {
"SubprocessEnvManager._take_step": {
"total": 3352.2454683749647,
"count": 325557,
"self": 51.12933249377738,
"children": {
"TorchPolicy.evaluate": {
"total": 3301.1161358811873,
"count": 646628,
"self": 3301.1161358811873
}
}
},
"workers": {
"total": 4.210227994659363,
"count": 325557,
"self": 0.0,
"children": {
"worker_root": {
"total": 11721.445844085025,
"count": 325557,
"is_parallel": true,
"self": 6097.572609526,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.004456181000023207,
"count": 2,
"is_parallel": true,
"self": 0.0011572050007089274,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0032989759993142798,
"count": 8,
"is_parallel": true,
"self": 0.0032989759993142798
}
}
},
"UnityEnvironment.step": {
"total": 0.08418834999974933,
"count": 1,
"is_parallel": true,
"self": 0.0006610949994865223,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0006977160001042648,
"count": 1,
"is_parallel": true,
"self": 0.0006977160001042648
},
"communicator.exchange": {
"total": 0.08099854900001446,
"count": 1,
"is_parallel": true,
"self": 0.08099854900001446
},
"steps_from_proto": {
"total": 0.0018309900001440838,
"count": 2,
"is_parallel": true,
"self": 0.0004084680003870744,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014225219997570093,
"count": 8,
"is_parallel": true,
"self": 0.0014225219997570093
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 5623.815688774028,
"count": 325556,
"is_parallel": true,
"self": 177.2967017169376,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 108.48498381085847,
"count": 325556,
"is_parallel": true,
"self": 108.48498381085847
},
"communicator.exchange": {
"total": 4808.189768435477,
"count": 325556,
"is_parallel": true,
"self": 4808.189768435477
},
"steps_from_proto": {
"total": 529.8442348107546,
"count": 651112,
"is_parallel": true,
"self": 109.20854704606609,
"children": {
"_process_rank_one_or_two_observation": {
"total": 420.63568776468856,
"count": 2604448,
"is_parallel": true,
"self": 420.63568776468856
}
}
}
}
},
"steps_from_proto": {
"total": 0.05754578499681884,
"count": 48,
"is_parallel": true,
"self": 0.011645412000689248,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.04590037299612959,
"count": 192,
"is_parallel": true,
"self": 0.04590037299612959
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 1911.9531908766567,
"count": 325557,
"self": 56.849382719127334,
"children": {
"process_trajectory": {
"total": 687.2321112195373,
"count": 325557,
"self": 683.3094099615414,
"children": {
"RLTrainer._checkpoint": {
"total": 3.922701257995868,
"count": 10,
"self": 3.922701257995868
}
}
},
"_update_policy": {
"total": 1167.871696937992,
"count": 230,
"self": 567.2716414120696,
"children": {
"TorchPOCAOptimizer.update": {
"total": 600.6000555259225,
"count": 6909,
"self": 600.6000555259225
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.776999852154404e-06,
"count": 1,
"self": 1.776999852154404e-06
},
"TrainerController._save_models": {
"total": 0.3213817729993025,
"count": 1,
"self": 0.016208342000027187,
"children": {
"RLTrainer._checkpoint": {
"total": 0.3051734309992753,
"count": 1,
"self": 0.3051734309992753
}
}
}
}
}
}
}