{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4067920446395874, "min": 1.4067853689193726, "max": 1.4296510219573975, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 70473.25, "min": 68201.796875, "max": 79200.515625, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 82.74539363484087, "min": 79.66559485530547, "max": 393.0625, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49399.0, "min": 49216.0, "max": 50312.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999919.0, "min": 49965.0, "max": 1999919.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999919.0, "min": 49965.0, "max": 1999919.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4877970218658447, "min": 0.06911392509937286, "max": 2.505692720413208, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1485.21484375, "min": 8.77746868133545, "max": 1511.541015625, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.827658890879134, "min": 1.942420170063109, "max": 3.9731475588800285, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2285.112357854843, "min": 246.68736159801483, "max": 2360.852304816246, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.827658890879134, "min": 1.942420170063109, "max": 3.9731475588800285, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2285.112357854843, "min": 246.68736159801483, "max": 2360.852304816246, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.0168516275667773, "min": 0.013052999078596864, "max": 0.01917462457786314, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.0505548827003319, "min": 0.02610599815719373, "max": 0.056180708234508825, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.05988596582578288, "min": 0.02060506163785855, "max": 0.06178859863430262, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.17965789747734864, "min": 0.0412101232757171, "max": 0.18092184625566005, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.668248777283336e-06, "min": 3.668248777283336e-06, "max": 0.0002953689015437, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.1004746331850008e-05, "min": 1.1004746331850008e-05, "max": 0.00084434716855095, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10122271666666667, "min": 0.10122271666666667, "max": 0.19845629999999997, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30366815, "min": 0.20758455, "max": 0.58144905, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 7.101356166666673e-05, "min": 7.101356166666673e-05, "max": 0.00492296937, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.0002130406850000002, "min": 0.0002130406850000002, "max": 0.014074307595, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1671105573", "python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1671107699" }, "total": 2125.481668978, "count": 1, "self": 0.3881110160004937, "children": { "run_training.setup": { "total": 0.10592237499997736, "count": 1, "self": 0.10592237499997736 }, "TrainerController.start_learning": { "total": 2124.9876355869997, "count": 1, "self": 3.614659176069381, "children": { "TrainerController._reset_env": { "total": 10.246105261000025, "count": 1, "self": 10.246105261000025 }, "TrainerController.advance": { "total": 2111.0160750019304, "count": 232775, "self": 3.849623986882307, "children": { "env_step": { "total": 1652.431737668068, "count": 232775, "self": 1389.374624187849, "children": { "SubprocessEnvManager._take_step": { "total": 260.5560644671199, "count": 232775, "self": 13.614857373106133, "children": { "TorchPolicy.evaluate": { "total": 246.94120709401375, "count": 222948, "self": 62.13880157205608, "children": { "TorchPolicy.sample_actions": { "total": 184.80240552195767, "count": 222948, "self": 184.80240552195767 } } } } }, "workers": { "total": 2.501049013099305, "count": 232775, "self": 0.0, "children": { "worker_root": { "total": 2117.603022251928, "count": 232775, "is_parallel": true, "self": 973.2808506069696, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0021021490000521226, "count": 1, "is_parallel": true, "self": 0.0003184160001410419, "children": { "_process_rank_one_or_two_observation": { "total": 0.0017837329999110807, "count": 2, "is_parallel": true, "self": 0.0017837329999110807 } } }, "UnityEnvironment.step": { "total": 0.02698912100004236, "count": 1, "is_parallel": true, "self": 0.0002904630001694386, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00022752899997158238, "count": 1, "is_parallel": true, "self": 0.00022752899997158238 }, "communicator.exchange": { "total": 0.025767768999912732, "count": 1, "is_parallel": true, "self": 0.025767768999912732 }, "steps_from_proto": { "total": 0.0007033599999886064, "count": 1, "is_parallel": true, "self": 0.00023073499983183865, "children": { "_process_rank_one_or_two_observation": { "total": 0.0004726250001567678, "count": 2, "is_parallel": true, "self": 0.0004726250001567678 } } } } } } }, "UnityEnvironment.step": { "total": 1144.3221716449584, "count": 232774, "is_parallel": true, "self": 33.95075012582083, "children": { "UnityEnvironment._generate_step_input": { "total": 73.57061031805006, "count": 232774, "is_parallel": true, "self": 73.57061031805006 }, "communicator.exchange": { "total": 946.6710108980604, "count": 232774, "is_parallel": true, "self": 946.6710108980604 }, "steps_from_proto": { "total": 90.12980030302697, "count": 232774, "is_parallel": true, "self": 37.1070635228383, "children": { "_process_rank_one_or_two_observation": { "total": 53.022736780188666, "count": 465548, "is_parallel": true, "self": 53.022736780188666 } } } } } } } } } } }, "trainer_advance": { "total": 454.7347133469799, "count": 232775, "self": 5.893178097013447, "children": { "process_trajectory": { "total": 144.08610869396534, "count": 232775, "self": 142.94394862396564, "children": { "RLTrainer._checkpoint": { "total": 1.1421600699997043, "count": 10, "self": 1.1421600699997043 } } }, "_update_policy": { "total": 304.7554265560011, "count": 97, "self": 252.05411180900455, "children": { "TorchPPOOptimizer.update": { "total": 52.70131474699656, "count": 2910, "self": 52.70131474699656 } } } } } } }, "trainer_threads": { "total": 9.559998943586834e-07, "count": 1, "self": 9.559998943586834e-07 }, "TrainerController._save_models": { "total": 0.11079519199984134, "count": 1, "self": 0.0019407079998927657, "children": { "RLTrainer._checkpoint": { "total": 0.10885448399994857, "count": 1, "self": 0.10885448399994857 } } } } } } }