{ "name": "root", "gauges": { "Agent.Policy.Entropy.mean": { "value": 1.3894017934799194, "min": 1.3879185914993286, "max": 1.4296987056732178, "count": 200 }, "Agent.Policy.Entropy.sum": { "value": 9736.927734375, "min": 6869.244140625, "max": 9918.16015625, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.mean": { "value": 0.9411764705882353, "min": 0.2777777777777778, "max": 0.9797979797979798, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.sum": { "value": 96.0, "min": 5.0, "max": 105.0, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.mean": { "value": 10.843137254901961, "min": 2.9444444444444446, "max": 44.8, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.sum": { "value": 1106.0, "min": 53.0, "max": 1344.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.mean": { "value": 0.0392156862745098, "min": 0.01282051282051282, "max": 0.5238095238095238, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.sum": { "value": 4.0, "min": 1.0, "max": 12.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": { "value": 48.73601814344818, "min": 10.470068295796713, "max": 59.34038262897067, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": { "value": 4971.073850631714, "min": 188.46122932434082, "max": 4971.073850631714, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": { "value": 3.7945638117836973, "min": 0.2459058202803135, "max": 5.265263577302297, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": { "value": 387.0455088019371, "min": 5.901739686727524, "max": 465.2266719341278, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": { "value": 1.1126193328230989, "min": 0.19778246538979666, "max": 1.985036908476441, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": { "value": 113.48717194795609, "min": 4.15343177318573, "max": 140.9568480849266, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": { "value": 0.1112619328732584, "min": 0.01977824668089549, "max": 0.19850369280687086, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": { "value": 11.348717153072357, "min": 0.41534318029880524, "max": 14.095684811472893, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": { "value": 65.04783301727444, "min": 19.413701603809994, "max": 78.60631868574355, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": { "value": 6634.878967761993, "min": 349.8026351928711, "max": 7465.602198660374, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.EpisodeLength.mean": { "value": 67.79411764705883, "min": 52.55555555555556, "max": 333.5, "count": 200 }, "Agent.Environment.EpisodeLength.sum": { "value": 6915.0, "min": 4863.0, "max": 7047.0, "count": 200 }, "Agent.Step.mean": { "value": 1199853.0, "min": 5673.0, "max": 1199853.0, "count": 200 }, "Agent.Step.sum": { "value": 1199853.0, "min": 5673.0, "max": 1199853.0, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.mean": { "value": 0.21435299515724182, "min": 0.013194466009736061, "max": 0.9914538860321045, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.sum": { "value": 21.435298919677734, "min": 0.39583396911621094, "max": 39.37849426269531, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.mean": { "value": 48.628662109375, "min": 0.05967150628566742, "max": 53.77576446533203, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.sum": { "value": 4862.8662109375, "min": 1.4917876720428467, "max": 5449.505859375, "count": 200 }, "Agent.Environment.CumulativeReward.mean": { "value": 100.98866849931888, "min": 29.839538896973764, "max": 104.83166097342968, "count": 200 }, "Agent.Environment.CumulativeReward.sum": { "value": 10098.866849931888, "min": 537.1117001455277, "max": 11197.802248001099, "count": 200 }, "Agent.Policy.CuriosityReward.mean": { "value": 0.16741907527670263, "min": 0.0, "max": 5.98461788041251, "count": 200 }, "Agent.Policy.CuriosityReward.sum": { "value": 16.741907527670264, "min": 0.0, "max": 209.46162581443787, "count": 200 }, "Agent.Policy.ExtrinsicReward.mean": { "value": 90.88980428959243, "min": 26.85558773132248, "max": 94.34849665284156, "count": 200 }, "Agent.Policy.ExtrinsicReward.sum": { "value": 9088.980428959243, "min": 483.40057916380465, "max": 10078.0221850276, "count": 200 }, "Agent.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.Losses.PolicyLoss.mean": { "value": 0.016831843919741612, "min": 0.015841431605319183, "max": 0.03535402244112144, "count": 138 }, "Agent.Losses.PolicyLoss.sum": { "value": 0.016831843919741612, "min": 0.015841431605319183, "max": 0.03535402244112144, "count": 138 }, "Agent.Losses.ValueLoss.mean": { "value": 111.00194835662842, "min": 13.606821960873074, "max": 143.2328265508016, "count": 138 }, "Agent.Losses.ValueLoss.sum": { "value": 111.00194835662842, "min": 13.606821960873074, "max": 143.2328265508016, "count": 138 }, "Agent.Policy.LearningRate.mean": { "value": 1.7334994225000254e-07, "min": 1.7334994225000254e-07, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.LearningRate.sum": { "value": 1.7334994225000254e-07, "min": 1.7334994225000254e-07, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.Epsilon.mean": { "value": 0.10005775, "min": 0.10005775, "max": 0.1992545, "count": 138 }, "Agent.Policy.Epsilon.sum": { "value": 0.10005775, "min": 0.10005775, "max": 0.1992545, "count": 138 }, "Agent.Policy.Beta.mean": { "value": 1.2881725000000045e-05, "min": 1.2881725000000045e-05, "max": 0.00496279955, "count": 138 }, "Agent.Policy.Beta.sum": { "value": 1.2881725000000045e-05, "min": 1.2881725000000045e-05, "max": 0.00496279955, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.mean": { "value": 0.026166158107419808, "min": 0.019183928923060495, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.sum": { "value": 0.026166158107419808, "min": 0.019183928923060495, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.mean": { "value": 2.3459423085053763, "min": 2.3459423085053763, "max": 3.353024572134018, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.sum": { "value": 2.3459423085053763, "min": 2.3459423085053763, "max": 3.353024572134018, "count": 138 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1716770949", "python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_1_task_3_run_id_0_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_1_task_3_run_id_0_train --base-port 5007", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.7.1+cu110", "numpy_version": "1.21.0", "end_time_seconds": "1716776065" }, "total": 5116.0386944, "count": 1, "self": 0.6011436999997386, "children": { "run_training.setup": { "total": 0.05958679999999994, "count": 1, "self": 0.05958679999999994 }, "TrainerController.start_learning": { "total": 5115.3779639, "count": 1, "self": 6.18105180020666, "children": { "TrainerController._reset_env": { "total": 2.346834, "count": 1, "self": 2.346834 }, "TrainerController.advance": { "total": 5106.6660246997935, "count": 403029, "self": 5.535118899771987, "children": { "env_step": { "total": 5101.1309058000215, "count": 403029, "self": 2223.429386500185, "children": { "SubprocessEnvManager._take_step": { "total": 2873.8766088999923, "count": 403029, "self": 11.38343330004136, "children": { "TorchPolicy.evaluate": { "total": 2862.493175599951, "count": 400351, "self": 2862.493175599951 } } }, "workers": { "total": 3.82491039984438, "count": 403029, "self": 0.0, "children": { "worker_root": { "total": 5105.519586199718, "count": 403029, "is_parallel": true, "self": 3133.7514020996114, "children": { "steps_from_proto": { "total": 0.007135700000000078, "count": 1, "is_parallel": true, "self": 0.00010679999999996248, "children": { "_process_maybe_compressed_observation": { "total": 0.006981499999999974, "count": 2, "is_parallel": true, "self": 3.49999999997852e-05, "children": { "_observation_to_np_array": { "total": 0.006946500000000189, "count": 3, "is_parallel": true, "self": 3.030000000014965e-05, "children": { "process_pixels": { "total": 0.006916200000000039, "count": 3, "is_parallel": true, "self": 0.0002423000000000286, "children": { "image_decompress": { "total": 0.00667390000000001, "count": 3, "is_parallel": true, "self": 0.00667390000000001 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 4.740000000014177e-05, "count": 2, "is_parallel": true, "self": 4.740000000014177e-05 } } }, "UnityEnvironment.step": { "total": 1971.7610484001066, "count": 403029, "is_parallel": true, "self": 19.534251700316418, "children": { "UnityEnvironment._generate_step_input": { "total": 19.679736499877272, "count": 403029, "is_parallel": true, "self": 19.679736499877272 }, "communicator.exchange": { "total": 1781.4259515998579, "count": 403029, "is_parallel": true, "self": 1781.4259515998579 }, "steps_from_proto": { "total": 151.12110860005515, "count": 403029, "is_parallel": true, "self": 30.234067699896826, "children": { "_process_maybe_compressed_observation": { "total": 108.4416234999839, "count": 806058, "is_parallel": true, "self": 8.572167100125085, "children": { "_observation_to_np_array": { "total": 99.86945639985882, "count": 1209741, "is_parallel": true, "self": 8.189591099849153, "children": { "process_pixels": { "total": 91.67986530000967, "count": 1209741, "is_parallel": true, "self": 42.656509400110764, "children": { "image_decompress": { "total": 49.0233558998989, "count": 1209741, "is_parallel": true, "self": 49.0233558998989 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 12.445417400174419, "count": 806058, "is_parallel": true, "self": 12.445417400174419 } } } } } } } } } } } } }, "trainer_threads": { "total": 2.7600000066740904e-05, "count": 1, "self": 2.7600000066740904e-05, "children": { "thread_root": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "trainer_advance": { "total": 5110.512504799872, "count": 250081, "is_parallel": true, "self": 7.453886899676945, "children": { "process_trajectory": { "total": 4139.222991800196, "count": 250081, "is_parallel": true, "self": 4138.781529700196, "children": { "RLTrainer._checkpoint": { "total": 0.44146209999962593, "count": 2, "is_parallel": true, "self": 0.44146209999962593 } } }, "_update_policy": { "total": 963.8356260999994, "count": 138, "is_parallel": true, "self": 633.5532193999976, "children": { "TorchPPOOptimizer.update": { "total": 330.28240670000184, "count": 3348, "is_parallel": true, "self": 330.28240670000184 } } } } } } } } }, "TrainerController._save_models": { "total": 0.18402580000019952, "count": 1, "self": 0.004897300000266114, "children": { "RLTrainer._checkpoint": { "total": 0.1791284999999334, "count": 1, "self": 0.1791284999999334 } } } } } } }