{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4029500484466553, "min": 1.4029500484466553, "max": 1.427134394645691, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 68240.890625, "min": 67824.8515625, "max": 79353.578125, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 117.40995260663507, "min": 84.11945392491468, "max": 381.4503816793893, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49547.0, "min": 48940.0, "max": 50368.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999946.0, "min": 49950.0, "max": 1999946.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999946.0, "min": 49950.0, "max": 1999946.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.306443691253662, "min": 0.07948636263608932, "max": 2.4858205318450928, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 973.3192138671875, "min": 10.333227157592773, "max": 1438.078125, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.4616845073575657, "min": 1.8123217739737951, "max": 4.105002243518829, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 1460.8308621048927, "min": 235.60183061659336, "max": 2225.898113846779, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.4616845073575657, "min": 1.8123217739737951, "max": 4.105002243518829, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 1460.8308621048927, "min": 235.60183061659336, "max": 2225.898113846779, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.016154403954796077, "min": 0.013882778619558342, "max": 0.02106150973607631, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.048463211864388234, "min": 0.027858269662829116, "max": 0.05514129457429438, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.04534555743965838, "min": 0.022708302984635036, "max": 0.06060358677059412, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.13603667231897515, "min": 0.04541660596927007, "max": 0.17140753430624803, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.3400488866833246e-06, "min": 3.3400488866833246e-06, "max": 0.00029532255155915007, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.0020146660049974e-05, "min": 1.0020146660049974e-05, "max": 0.0008440612686462497, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10111331666666669, "min": 0.10111331666666669, "max": 0.19844085, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.3033399500000001, "min": 0.2073955, "max": 0.58135375, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 6.555450166666654e-05, "min": 6.555450166666654e-05, "max": 0.0049221984150000005, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00019666350499999963, "min": 0.00019666350499999963, "max": 0.014069552125000004, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1670899706", "python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1670901918" }, "total": 2211.9882652610004, "count": 1, "self": 0.4348890110004504, "children": { "run_training.setup": { "total": 0.10711237100008475, "count": 1, "self": 0.10711237100008475 }, "TrainerController.start_learning": { "total": 2211.446263879, "count": 1, "self": 3.9073637871310893, "children": { "TrainerController._reset_env": { "total": 10.172514633999981, "count": 1, "self": 10.172514633999981 }, "TrainerController.advance": { "total": 2197.250240443869, "count": 232421, "self": 3.9935037146760806, "children": { "env_step": { "total": 1722.4658394911257, "count": 232421, "self": 1445.503340941234, "children": { "SubprocessEnvManager._take_step": { "total": 274.3513423609877, "count": 232421, "self": 14.228028341023105, "children": { "TorchPolicy.evaluate": { "total": 260.1233140199646, "count": 222993, "self": 65.34839569591259, "children": { "TorchPolicy.sample_actions": { "total": 194.774918324052, "count": 222993, "self": 194.774918324052 } } } } }, "workers": { "total": 2.611156188904147, "count": 232421, "self": 0.0, "children": { "worker_root": { "total": 2203.685773803074, "count": 232421, "is_parallel": true, "self": 1018.9052431620112, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0019430820000252425, "count": 1, "is_parallel": true, "self": 0.000311869000029219, "children": { "_process_rank_one_or_two_observation": { "total": 0.0016312129999960234, "count": 2, "is_parallel": true, "self": 0.0016312129999960234 } } }, "UnityEnvironment.step": { "total": 0.02809839400003966, "count": 1, "is_parallel": true, "self": 0.0002744500000062544, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0001773080000475602, "count": 1, "is_parallel": true, "self": 0.0001773080000475602 }, "communicator.exchange": { "total": 0.0269492289999107, "count": 1, "is_parallel": true, "self": 0.0269492289999107 }, "steps_from_proto": { "total": 0.0006974070000751453, "count": 1, "is_parallel": true, "self": 0.00023462100000415376, "children": { "_process_rank_one_or_two_observation": { "total": 0.0004627860000709916, "count": 2, "is_parallel": true, "self": 0.0004627860000709916 } } } } } } }, "UnityEnvironment.step": { "total": 1184.7805306410628, "count": 232420, "is_parallel": true, "self": 34.19085954609341, "children": { "UnityEnvironment._generate_step_input": { "total": 74.99743670295618, "count": 232420, "is_parallel": true, "self": 74.99743670295618 }, "communicator.exchange": { "total": 982.9845625260191, "count": 232420, "is_parallel": true, "self": 982.9845625260191 }, "steps_from_proto": { "total": 92.607671865994, "count": 232420, "is_parallel": true, "self": 37.830659432086236, "children": { "_process_rank_one_or_two_observation": { "total": 54.77701243390777, "count": 464840, "is_parallel": true, "self": 54.77701243390777 } } } } } } } } } } }, "trainer_advance": { "total": 470.7908972380675, "count": 232421, "self": 6.257566197091251, "children": { "process_trajectory": { "total": 149.21605765297465, "count": 232421, "self": 148.74709216497433, "children": { "RLTrainer._checkpoint": { "total": 0.4689654880003218, "count": 4, "self": 0.4689654880003218 } } }, "_update_policy": { "total": 315.3172733880016, "count": 97, "self": 260.94258537999417, "children": { "TorchPPOOptimizer.update": { "total": 54.37468800800741, "count": 2910, "self": 54.37468800800741 } } } } } } }, "trainer_threads": { "total": 8.729998626222368e-07, "count": 1, "self": 8.729998626222368e-07 }, "TrainerController._save_models": { "total": 0.11614414099994974, "count": 1, "self": 0.0021073560001241276, "children": { "RLTrainer._checkpoint": { "total": 0.11403678499982561, "count": 1, "self": 0.11403678499982561 } } } } } } }