ppo-Huggy-01 / run_logs /timers.json
ibadrehman's picture
Huggy
3c2b7e6
raw
history blame
17.9 kB
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4067920446395874,
"min": 1.4067853689193726,
"max": 1.4296510219573975,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70473.25,
"min": 68201.796875,
"max": 79200.515625,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 82.74539363484087,
"min": 79.66559485530547,
"max": 393.0625,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49399.0,
"min": 49216.0,
"max": 50312.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999919.0,
"min": 49965.0,
"max": 1999919.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999919.0,
"min": 49965.0,
"max": 1999919.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4877970218658447,
"min": 0.06911392509937286,
"max": 2.505692720413208,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1485.21484375,
"min": 8.77746868133545,
"max": 1511.541015625,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.827658890879134,
"min": 1.942420170063109,
"max": 3.9731475588800285,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2285.112357854843,
"min": 246.68736159801483,
"max": 2360.852304816246,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.827658890879134,
"min": 1.942420170063109,
"max": 3.9731475588800285,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2285.112357854843,
"min": 246.68736159801483,
"max": 2360.852304816246,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.0168516275667773,
"min": 0.013052999078596864,
"max": 0.01917462457786314,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.0505548827003319,
"min": 0.02610599815719373,
"max": 0.056180708234508825,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.05988596582578288,
"min": 0.02060506163785855,
"max": 0.06178859863430262,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.17965789747734864,
"min": 0.0412101232757171,
"max": 0.18092184625566005,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.668248777283336e-06,
"min": 3.668248777283336e-06,
"max": 0.0002953689015437,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.1004746331850008e-05,
"min": 1.1004746331850008e-05,
"max": 0.00084434716855095,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10122271666666667,
"min": 0.10122271666666667,
"max": 0.19845629999999997,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30366815,
"min": 0.20758455,
"max": 0.58144905,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 7.101356166666673e-05,
"min": 7.101356166666673e-05,
"max": 0.00492296937,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0002130406850000002,
"min": 0.0002130406850000002,
"max": 0.014074307595,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1671105573",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1671107699"
},
"total": 2125.481668978,
"count": 1,
"self": 0.3881110160004937,
"children": {
"run_training.setup": {
"total": 0.10592237499997736,
"count": 1,
"self": 0.10592237499997736
},
"TrainerController.start_learning": {
"total": 2124.9876355869997,
"count": 1,
"self": 3.614659176069381,
"children": {
"TrainerController._reset_env": {
"total": 10.246105261000025,
"count": 1,
"self": 10.246105261000025
},
"TrainerController.advance": {
"total": 2111.0160750019304,
"count": 232775,
"self": 3.849623986882307,
"children": {
"env_step": {
"total": 1652.431737668068,
"count": 232775,
"self": 1389.374624187849,
"children": {
"SubprocessEnvManager._take_step": {
"total": 260.5560644671199,
"count": 232775,
"self": 13.614857373106133,
"children": {
"TorchPolicy.evaluate": {
"total": 246.94120709401375,
"count": 222948,
"self": 62.13880157205608,
"children": {
"TorchPolicy.sample_actions": {
"total": 184.80240552195767,
"count": 222948,
"self": 184.80240552195767
}
}
}
}
},
"workers": {
"total": 2.501049013099305,
"count": 232775,
"self": 0.0,
"children": {
"worker_root": {
"total": 2117.603022251928,
"count": 232775,
"is_parallel": true,
"self": 973.2808506069696,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0021021490000521226,
"count": 1,
"is_parallel": true,
"self": 0.0003184160001410419,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0017837329999110807,
"count": 2,
"is_parallel": true,
"self": 0.0017837329999110807
}
}
},
"UnityEnvironment.step": {
"total": 0.02698912100004236,
"count": 1,
"is_parallel": true,
"self": 0.0002904630001694386,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00022752899997158238,
"count": 1,
"is_parallel": true,
"self": 0.00022752899997158238
},
"communicator.exchange": {
"total": 0.025767768999912732,
"count": 1,
"is_parallel": true,
"self": 0.025767768999912732
},
"steps_from_proto": {
"total": 0.0007033599999886064,
"count": 1,
"is_parallel": true,
"self": 0.00023073499983183865,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0004726250001567678,
"count": 2,
"is_parallel": true,
"self": 0.0004726250001567678
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1144.3221716449584,
"count": 232774,
"is_parallel": true,
"self": 33.95075012582083,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 73.57061031805006,
"count": 232774,
"is_parallel": true,
"self": 73.57061031805006
},
"communicator.exchange": {
"total": 946.6710108980604,
"count": 232774,
"is_parallel": true,
"self": 946.6710108980604
},
"steps_from_proto": {
"total": 90.12980030302697,
"count": 232774,
"is_parallel": true,
"self": 37.1070635228383,
"children": {
"_process_rank_one_or_two_observation": {
"total": 53.022736780188666,
"count": 465548,
"is_parallel": true,
"self": 53.022736780188666
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 454.7347133469799,
"count": 232775,
"self": 5.893178097013447,
"children": {
"process_trajectory": {
"total": 144.08610869396534,
"count": 232775,
"self": 142.94394862396564,
"children": {
"RLTrainer._checkpoint": {
"total": 1.1421600699997043,
"count": 10,
"self": 1.1421600699997043
}
}
},
"_update_policy": {
"total": 304.7554265560011,
"count": 97,
"self": 252.05411180900455,
"children": {
"TorchPPOOptimizer.update": {
"total": 52.70131474699656,
"count": 2910,
"self": 52.70131474699656
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.559998943586834e-07,
"count": 1,
"self": 9.559998943586834e-07
},
"TrainerController._save_models": {
"total": 0.11079519199984134,
"count": 1,
"self": 0.0019407079998927657,
"children": {
"RLTrainer._checkpoint": {
"total": 0.10885448399994857,
"count": 1,
"self": 0.10885448399994857
}
}
}
}
}
}
}