philippds's picture
Upload 8 files
5b3338f verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.9662578105926514,
"min": 3.8902809619903564,
"max": 4.3802947998046875,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 19525.88671875,
"min": 3942.199951171875,
"max": 31338.61328125,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8888887448443307,
"min": 0.5500000417232513,
"max": 0.8999999136560493,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 15.999997407197952,
"min": 7.5,
"max": 32.299995206296444,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 1026.1174051894081,
"min": 1.2436208843460514,
"max": 2813.268530951606,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 18470.113293409348,
"min": 11.192587959114462,
"max": 36061.0396194458,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 0.2600799028037323,
"min": 0.013587533599800534,
"max": 4.705731664680773,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 4.681438250467181,
"min": 0.1222878023982048,
"max": 84.7031699642539,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 57.71533436245389,
"min": 0.2929872473080953,
"max": 174.170406765408,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 1038.87601852417,
"min": 2.6368852257728577,
"max": 2244.8185958862305,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 35.585473120212555,
"min": 0.10056810929543442,
"max": 78.23635286755032,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 640.538516163826,
"min": 0.9051129836589098,
"max": 1107.8064243793488,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 273.0,
"min": 16.0,
"max": 455.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 4914.0,
"min": 144.0,
"max": 7695.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449877.0,
"min": 4381.0,
"max": 449877.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449877.0,
"min": 4381.0,
"max": 449877.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": -0.0028904082719236612,
"min": -0.0634576678276062,
"max": 0.12072662264108658,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": -0.04624653235077858,
"min": -1.269153356552124,
"max": 3.13889217376709,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 246.62399291992188,
"min": 6.235428333282471,
"max": 357.42041015625,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 3945.98388671875,
"min": 162.1211395263672,
"max": 11794.8740234375,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 930.314374923706,
"min": 193.45669285456339,
"max": 2241.3963661193848,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 14885.029998779297,
"min": 5150.876541137695,
"max": 35862.341857910156,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.015021562692709267,
"min": 0.009544561451841746,
"max": 0.22255233503304994,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.24034500308334827,
"min": 0.19084811629727483,
"max": 5.786360710859299,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 930.314374923706,
"min": 193.45669285456339,
"max": 2241.3963661193848,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 14885.029998779297,
"min": 5150.876541137695,
"max": 35862.341857910156,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.0636149788009529,
"min": 0.05675364297533792,
"max": 0.08363459916303359,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.1272299576019058,
"min": 0.05675364297533792,
"max": 0.16726919832606718,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 3978.190133520818,
"min": 274.94069978560526,
"max": 13049.90203814956,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 7956.380267041636,
"min": 274.94069978560526,
"max": 26099.80407629912,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.8030993989999881e-06,
"min": 1.8030993989999881e-06,
"max": 0.00029819400060199984,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 3.6061987979999762e-06,
"min": 3.6061987979999762e-06,
"max": 0.0005861940046020002,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.10060100000000004,
"min": 0.10060100000000004,
"max": 0.19939800000000005,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.20120200000000008,
"min": 0.10181000000000004,
"max": 0.3953980000000001,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 7.003989999999961e-05,
"min": 7.003989999999961e-05,
"max": 0.009939860200000004,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00014007979999999923,
"min": 0.00014007979999999923,
"max": 0.019540260200000006,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.002147874424402681,
"min": 0.002037737639971398,
"max": 0.38484019738814185,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.004295748848805362,
"min": 0.002037737639971398,
"max": 0.38484019738814185,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.9688596760422605,
"min": 3.8964975935842854,
"max": 4.389180887313116,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.937719352084521,
"min": 3.911028209659788,
"max": 8.685674510777943,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716768588",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_5_task_2_run_id_2_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_5_task_2_run_id_2_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716770535"
},
"total": 1947.7930261,
"count": 1,
"self": 0.2078475000000708,
"children": {
"run_training.setup": {
"total": 0.05632779999999993,
"count": 1,
"self": 0.05632779999999993
},
"TrainerController.start_learning": {
"total": 1947.5288507999999,
"count": 1,
"self": 1.0826197000017146,
"children": {
"TrainerController._reset_env": {
"total": 2.3188974,
"count": 1,
"self": 2.3188974
},
"TrainerController.advance": {
"total": 1944.0281442999983,
"count": 50091,
"self": 0.9824415999921712,
"children": {
"env_step": {
"total": 1943.0457027000061,
"count": 50091,
"self": 1532.7661559999997,
"children": {
"SubprocessEnvManager._take_step": {
"total": 409.83407730001545,
"count": 50091,
"self": 1.8083528000054798,
"children": {
"TorchPolicy.evaluate": {
"total": 408.02572450000997,
"count": 50091,
"self": 408.02572450000997
}
}
},
"workers": {
"total": 0.4454693999908712,
"count": 50091,
"self": 0.0,
"children": {
"worker_root": {
"total": 1944.6120570000066,
"count": 50091,
"is_parallel": true,
"self": 472.48266330000774,
"children": {
"steps_from_proto": {
"total": 0.0002790999999999766,
"count": 1,
"is_parallel": true,
"self": 0.0001265000000001404,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0001525999999998362,
"count": 2,
"is_parallel": true,
"self": 0.0001525999999998362
}
}
},
"UnityEnvironment.step": {
"total": 1472.129114599999,
"count": 50091,
"is_parallel": true,
"self": 3.5904693000172756,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 4.666798800006584,
"count": 50091,
"is_parallel": true,
"self": 4.666798800006584
},
"communicator.exchange": {
"total": 1454.0078471999661,
"count": 50091,
"is_parallel": true,
"self": 1454.0078471999661
},
"steps_from_proto": {
"total": 9.863999300009022,
"count": 50091,
"is_parallel": true,
"self": 5.160215900029229,
"children": {
"_process_rank_one_or_two_observation": {
"total": 4.703783399979793,
"count": 100182,
"is_parallel": true,
"self": 4.703783399979793
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.189999986512703e-05,
"count": 1,
"self": 2.189999986512703e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1944.1581723999896,
"count": 89380,
"is_parallel": true,
"self": 2.8366435999580517,
"children": {
"process_trajectory": {
"total": 1417.463181500031,
"count": 89380,
"is_parallel": true,
"self": 1417.463181500031
},
"_update_policy": {
"total": 523.8583473000007,
"count": 143,
"is_parallel": true,
"self": 254.35360870000864,
"children": {
"TorchPPOOptimizer.update": {
"total": 269.50473859999204,
"count": 10296,
"is_parallel": true,
"self": 269.50473859999204
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.09916750000002139,
"count": 1,
"self": 0.014539000000013402,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08462850000000799,
"count": 1,
"self": 0.08462850000000799
}
}
}
}
}
}
}