{ "name": "root", "gauges": { "Agent.Policy.Entropy.mean": { "value": 1.3848451375961304, "min": 1.3848425149917603, "max": 1.4281865358352661, "count": 200 }, "Agent.Policy.Entropy.sum": { "value": 9721.61328125, "min": 6707.294921875, "max": 9935.82421875, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.mean": { "value": 0.9130434782608695, "min": 0.3333333333333333, "max": 0.991869918699187, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.sum": { "value": 63.0, "min": 5.0, "max": 122.0, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.mean": { "value": 11.057971014492754, "min": 2.2, "max": 44.8, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.sum": { "value": 763.0, "min": 33.0, "max": 1486.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.mean": { "value": 0.07246376811594203, "min": 0.008130081300813009, "max": 0.5333333333333333, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.sum": { "value": 5.0, "min": 1.0, "max": 12.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": { "value": 52.58073242850926, "min": 12.207405726114908, "max": 61.72031497955322, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": { "value": 3628.0705375671387, "min": 250.79424285888672, "max": 5155.910724639893, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": { "value": 3.26475090150168, "min": 0.07396268049875895, "max": 6.10832218436355, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": { "value": 225.2678122036159, "min": 1.1094402074813843, "max": 572.8081967830658, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": { "value": 1.3341534353684688, "min": 0.2047844628492991, "max": 2.016074032584826, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": { "value": 92.05658704042435, "min": 3.5308263301849365, "max": 159.98608255386353, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": { "value": 0.13341534483260004, "min": 0.020478446036577225, "max": 0.20160740353167056, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": { "value": 9.205658793449402, "min": 0.3530826270580292, "max": 15.998608238995075, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": { "value": 65.95886754125789, "min": 13.084621217515734, "max": 77.89113126595815, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": { "value": 4551.161860346794, "min": 235.5231819152832, "max": 8497.22591650486, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.EpisodeLength.mean": { "value": 100.73913043478261, "min": 45.36585365853659, "max": 399.0, "count": 200 }, "Agent.Environment.EpisodeLength.sum": { "value": 6951.0, "min": 4746.0, "max": 7038.0, "count": 200 }, "Agent.Step.mean": { "value": 1199844.0, "min": 5673.0, "max": 1199844.0, "count": 200 }, "Agent.Step.sum": { "value": 1199844.0, "min": 5673.0, "max": 1199844.0, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.mean": { "value": 0.24977338314056396, "min": 0.013205585069954395, "max": 1.064765453338623, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.sum": { "value": 17.234363555908203, "min": 0.3961675465106964, "max": 34.85499572753906, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.mean": { "value": 50.36190414428711, "min": 0.05967150628566742, "max": 53.74601364135742, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.sum": { "value": 3474.971435546875, "min": 1.4917876720428467, "max": 5976.26953125, "count": 200 }, "Agent.Environment.CumulativeReward.mean": { "value": 100.01051464957604, "min": 33.38887961308161, "max": 106.3299504097062, "count": 200 }, "Agent.Environment.CumulativeReward.sum": { "value": 6900.725510820746, "min": 500.8331941962242, "max": 12919.049165844917, "count": 200 }, "Agent.Policy.CuriosityReward.mean": { "value": 0.20973035077685895, "min": 0.0, "max": 5.984334189551217, "count": 200 }, "Agent.Policy.CuriosityReward.sum": { "value": 14.471394203603268, "min": 0.0, "max": 209.4516966342926, "count": 200 }, "Agent.Policy.ExtrinsicReward.mean": { "value": 90.00946634943071, "min": 30.04999473889669, "max": 95.69695790969965, "count": 200 }, "Agent.Policy.ExtrinsicReward.sum": { "value": 6210.653178110719, "min": 450.7499210834503, "max": 11627.144557833672, "count": 200 }, "Agent.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.Losses.PolicyLoss.mean": { "value": 0.026116722729057074, "min": 0.014573930141826471, "max": 0.03793451966096958, "count": 138 }, "Agent.Losses.PolicyLoss.sum": { "value": 0.026116722729057074, "min": 0.014573930141826471, "max": 0.03793451966096958, "count": 138 }, "Agent.Losses.ValueLoss.mean": { "value": 97.28924910227458, "min": 22.247718532880146, "max": 144.694242477417, "count": 138 }, "Agent.Losses.ValueLoss.sum": { "value": 97.28924910227458, "min": 22.247718532880146, "max": 144.694242477417, "count": 138 }, "Agent.Policy.LearningRate.mean": { "value": 8.265997244999856e-07, "min": 8.265997244999856e-07, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.LearningRate.sum": { "value": 8.265997244999856e-07, "min": 8.265997244999856e-07, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.Epsilon.mean": { "value": 0.1002755, "min": 0.1002755, "max": 0.1992545, "count": 138 }, "Agent.Policy.Epsilon.sum": { "value": 0.1002755, "min": 0.1002755, "max": 0.1992545, "count": 138 }, "Agent.Policy.Beta.mean": { "value": 2.3747449999999762e-05, "min": 2.3747449999999762e-05, "max": 0.00496279955, "count": 138 }, "Agent.Policy.Beta.sum": { "value": 2.3747449999999762e-05, "min": 2.3747449999999762e-05, "max": 0.00496279955, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.mean": { "value": 0.025772807653993368, "min": 0.018984782819946606, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.sum": { "value": 0.025772807653993368, "min": 0.018984782819946606, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.mean": { "value": 2.3072385787963867, "min": 2.2800092498461404, "max": 3.353024572134018, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.sum": { "value": 2.3072385787963867, "min": 2.2800092498461404, "max": 3.353024572134018, "count": 138 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1717702123", "python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_7_task_3_run_id_2_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_7_task_3_run_id_2_train --base-port 5007", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.7.1+cu110", "numpy_version": "1.21.0", "end_time_seconds": "1717706411" }, "total": 4288.1489367, "count": 1, "self": 0.49392859999989014, "children": { "run_training.setup": { "total": 0.04943300000000006, "count": 1, "self": 0.04943300000000006 }, "TrainerController.start_learning": { "total": 4287.6055751, "count": 1, "self": 5.63005070006875, "children": { "TrainerController._reset_env": { "total": 2.0493757, "count": 1, "self": 2.0493757 }, "TrainerController.advance": { "total": 4279.68292289993, "count": 403195, "self": 5.4479560000054335, "children": { "env_step": { "total": 4274.234966899925, "count": 403195, "self": 2056.9615485001286, "children": { "SubprocessEnvManager._take_step": { "total": 2213.2936437999447, "count": 403195, "self": 11.579975600094258, "children": { "TorchPolicy.evaluate": { "total": 2201.7136681998504, "count": 400349, "self": 2201.7136681998504 } } }, "workers": { "total": 3.9797745998515874, "count": 403195, "self": 0.0, "children": { "worker_root": { "total": 4278.062895000038, "count": 403195, "is_parallel": true, "self": 2467.8336563999337, "children": { "steps_from_proto": { "total": 0.006490000000000107, "count": 1, "is_parallel": true, "self": 0.00010290000000034993, "children": { "_process_maybe_compressed_observation": { "total": 0.006341599999999836, "count": 2, "is_parallel": true, "self": 3.299999999950565e-05, "children": { "_observation_to_np_array": { "total": 0.006308600000000331, "count": 3, "is_parallel": true, "self": 2.990000000013815e-05, "children": { "process_pixels": { "total": 0.0062787000000001925, "count": 3, "is_parallel": true, "self": 0.00024170000000034442, "children": { "image_decompress": { "total": 0.006036999999999848, "count": 3, "is_parallel": true, "self": 0.006036999999999848 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 4.5499999999920604e-05, "count": 2, "is_parallel": true, "self": 4.5499999999920604e-05 } } }, "UnityEnvironment.step": { "total": 1810.2227486001043, "count": 403195, "is_parallel": true, "self": 22.132137400104284, "children": { "UnityEnvironment._generate_step_input": { "total": 22.734860300051785, "count": 403195, "is_parallel": true, "self": 22.734860300051785 }, "communicator.exchange": { "total": 1589.9010269999662, "count": 403195, "is_parallel": true, "self": 1589.9010269999662 }, "steps_from_proto": { "total": 175.45472389998184, "count": 403195, "is_parallel": true, "self": 34.780302600093904, "children": { "_process_maybe_compressed_observation": { "total": 125.5690601001191, "count": 806390, "is_parallel": true, "self": 10.114038300169184, "children": { "_observation_to_np_array": { "total": 115.45502179994992, "count": 1210191, "is_parallel": true, "self": 9.955208899894956, "children": { "process_pixels": { "total": 105.49981290005496, "count": 1210191, "is_parallel": true, "self": 49.71157090014538, "children": { "image_decompress": { "total": 55.78824199990958, "count": 1210191, "is_parallel": true, "self": 55.78824199990958 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 15.105361199768835, "count": 806390, "is_parallel": true, "self": 15.105361199768835 } } } } } } } } } } } } }, "trainer_threads": { "total": 3.04000004689442e-05, "count": 1, "self": 3.04000004689442e-05, "children": { "thread_root": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "trainer_advance": { "total": 4283.552069900078, "count": 208577, "is_parallel": true, "self": 4.947188199941593, "children": { "process_trajectory": { "total": 3470.770895900137, "count": 208577, "is_parallel": true, "self": 3470.3206993001368, "children": { "RLTrainer._checkpoint": { "total": 0.4501966000000266, "count": 2, "is_parallel": true, "self": 0.4501966000000266 } } }, "_update_policy": { "total": 807.8339857999997, "count": 138, "is_parallel": true, "self": 541.1189284, "children": { "TorchPPOOptimizer.update": { "total": 266.7150573999998, "count": 3354, "is_parallel": true, "self": 266.7150573999998 } } } } } } } } }, "TrainerController._save_models": { "total": 0.24319540000033157, "count": 1, "self": 0.007341900000028545, "children": { "RLTrainer._checkpoint": { "total": 0.23585350000030303, "count": 1, "self": 0.23585350000030303 } } } } } } }