{ "name": "root", "gauges": { "Agent.Policy.Entropy.mean": { "value": 1.461629033088684, "min": 1.4189385175704956, "max": 1.4624531269073486, "count": 200 }, "Agent.Policy.Entropy.sum": { "value": 8804.853515625, "min": 7319.8828125, "max": 10341.7890625, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.mean": { "value": 0.05555555555555555, "min": 0.0, "max": 394.8666666666667, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.sum": { "value": 1.0, "min": 0.0, "max": 5967.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.mean": { "value": 0.7222222222222222, "min": 0.3888888888888889, "max": 0.7333333333333333, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.sum": { "value": 13.0, "min": 6.0, "max": 13.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.EpisodeLength.mean": { "value": 335.0, "min": 284.25, "max": 399.0, "count": 200 }, "Agent.Environment.EpisodeLength.sum": { "value": 6030.0, "min": 5004.0, "max": 7101.0, "count": 200 }, "Agent.Step.mean": { "value": 1199702.0, "min": 5600.0, "max": 1199702.0, "count": 200 }, "Agent.Step.sum": { "value": 1199702.0, "min": 5600.0, "max": 1199702.0, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.mean": { "value": 0.12799793481826782, "min": 0.029348647221922874, "max": 1.0539227724075317, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.sum": { "value": 2.3039627075195312, "min": 0.44022971391677856, "max": 15.80884075164795, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.mean": { "value": 0.8996142148971558, "min": -0.19201916456222534, "max": 2.7600245475769043, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.sum": { "value": 16.193056106567383, "min": -3.0723066329956055, "max": 55.20048904418945, "count": 200 }, "Agent.Environment.CumulativeReward.mean": { "value": 4.716027929344111, "min": -1.1326333165168763, "max": 32.35911449293295, "count": 200 }, "Agent.Environment.CumulativeReward.sum": { "value": 84.888502728194, "min": -16.989499747753143, "max": 582.4640608727932, "count": 200 }, "Agent.Policy.CuriosityReward.mean": { "value": 0.4482337481652697, "min": 0.0, "max": 14.208261092503866, "count": 200 }, "Agent.Policy.CuriosityReward.sum": { "value": 8.068207466974854, "min": 0.0, "max": 255.74869966506958, "count": 200 }, "Agent.Policy.ExtrinsicReward.mean": { "value": 4.244424697839552, "min": -1.0193702220916747, "max": 29.12319684235586, "count": 200 }, "Agent.Policy.ExtrinsicReward.sum": { "value": 76.39964456111193, "min": -15.290553331375122, "max": 524.2175431624055, "count": 200 }, "Agent.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.Losses.PolicyLoss.mean": { "value": 0.02185002815288802, "min": 0.01476107951020822, "max": 0.03155452210921794, "count": 140 }, "Agent.Losses.PolicyLoss.sum": { "value": 0.02185002815288802, "min": 0.01476107951020822, "max": 0.03155452210921794, "count": 140 }, "Agent.Losses.ValueLoss.mean": { "value": 11.1973377764225, "min": 0.0014995217449419822, "max": 22.843181212743122, "count": 140 }, "Agent.Losses.ValueLoss.sum": { "value": 11.1973377764225, "min": 0.0014995217449419822, "max": 22.843181212743122, "count": 140 }, "Agent.Policy.LearningRate.mean": { "value": 5.745998085e-07, "min": 5.745998085e-07, "max": 0.0002979000007, "count": 140 }, "Agent.Policy.LearningRate.sum": { "value": 5.745998085e-07, "min": 5.745998085e-07, "max": 0.0002979000007, "count": 140 }, "Agent.Policy.Epsilon.mean": { "value": 0.1001915, "min": 0.1001915, "max": 0.1993, "count": 140 }, "Agent.Policy.Epsilon.sum": { "value": 0.1001915, "min": 0.1001915, "max": 0.1993, "count": 140 }, "Agent.Policy.Beta.mean": { "value": 1.9555850000000004e-05, "min": 1.9555850000000004e-05, "max": 0.00496507, "count": 140 }, "Agent.Policy.Beta.sum": { "value": 1.9555850000000004e-05, "min": 1.9555850000000004e-05, "max": 0.00496507, "count": 140 }, "Agent.Losses.CuriosityForwardLoss.mean": { "value": 0.013372027858470878, "min": 0.00993607259200265, "max": 0.5835270757476488, "count": 140 }, "Agent.Losses.CuriosityForwardLoss.sum": { "value": 0.013372027858470878, "min": 0.00993607259200265, "max": 0.5835270757476488, "count": 140 }, "Agent.Losses.CuriosityInverseLoss.mean": { "value": 2.372185468673706, "min": 2.317383666833242, "max": 3.3108297189076743, "count": 140 }, "Agent.Losses.CuriosityInverseLoss.sum": { "value": 2.372185468673706, "min": 2.317383666833242, "max": 3.3108297189076743, "count": 140 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1717759085", "python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_9_task_2_run_id_2_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_9_task_2_run_id_2_train --base-port 5007", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.7.1+cu110", "numpy_version": "1.21.0", "end_time_seconds": "1717762680" }, "total": 3594.6521040000002, "count": 1, "self": 0.2845319000002746, "children": { "run_training.setup": { "total": 0.05085230000000007, "count": 1, "self": 0.05085230000000007 }, "TrainerController.start_learning": { "total": 3594.3167198, "count": 1, "self": 4.735316700025123, "children": { "TrainerController._reset_env": { "total": 2.0784781000000003, "count": 1, "self": 2.0784781000000003 }, "TrainerController.advance": { "total": 3587.337379899975, "count": 400923, "self": 4.43875570006503, "children": { "env_step": { "total": 3582.89862419991, "count": 400923, "self": 1560.0457609998743, "children": { "SubprocessEnvManager._take_step": { "total": 2019.7551650000885, "count": 400923, "self": 10.148434200125394, "children": { "TorchPolicy.evaluate": { "total": 2009.606730799963, "count": 400035, "self": 2009.606730799963 } } }, "workers": { "total": 3.0976981999472724, "count": 400923, "self": 0.0, "children": { "worker_root": { "total": 3587.713058600138, "count": 400923, "is_parallel": true, "self": 2206.219591200131, "children": { "steps_from_proto": { "total": 0.006369700000000034, "count": 1, "is_parallel": true, "self": 0.00010019999999966167, "children": { "_process_maybe_compressed_observation": { "total": 0.006224900000000089, "count": 2, "is_parallel": true, "self": 2.750000000029118e-05, "children": { "_observation_to_np_array": { "total": 0.0061973999999997975, "count": 3, "is_parallel": true, "self": 2.8899999999776327e-05, "children": { "process_pixels": { "total": 0.006168500000000021, "count": 3, "is_parallel": true, "self": 0.0002241999999998967, "children": { "image_decompress": { "total": 0.0059443000000001245, "count": 3, "is_parallel": true, "self": 0.0059443000000001245 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 4.46000000002833e-05, "count": 2, "is_parallel": true, "self": 4.46000000002833e-05 } } }, "UnityEnvironment.step": { "total": 1381.4870977000069, "count": 400923, "is_parallel": true, "self": 15.452664199902756, "children": { "UnityEnvironment._generate_step_input": { "total": 18.870952999977682, "count": 400923, "is_parallel": true, "self": 18.870952999977682 }, "communicator.exchange": { "total": 1219.6390342000182, "count": 400923, "is_parallel": true, "self": 1219.6390342000182 }, "steps_from_proto": { "total": 127.52444630010842, "count": 400923, "is_parallel": true, "self": 25.686103800257825, "children": { "_process_maybe_compressed_observation": { "total": 91.04513479994904, "count": 801846, "is_parallel": true, "self": 7.155549800130132, "children": { "_observation_to_np_array": { "total": 83.88958499981891, "count": 1203327, "is_parallel": true, "self": 7.087501699546948, "children": { "process_pixels": { "total": 76.80208330027196, "count": 1203327, "is_parallel": true, "self": 36.39249940035934, "children": { "image_decompress": { "total": 40.40958389991262, "count": 1203327, "is_parallel": true, "self": 40.40958389991262 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 10.79320769990155, "count": 801846, "is_parallel": true, "self": 10.79320769990155 } } } } } } } } } } } } }, "trainer_threads": { "total": 3.819999983534217e-05, "count": 1, "self": 3.819999983534217e-05, "children": { "thread_root": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "trainer_advance": { "total": 3590.015360799979, "count": 173307, "is_parallel": true, "self": 5.554849099913099, "children": { "process_trajectory": { "total": 2830.8935339000654, "count": 173307, "is_parallel": true, "self": 2830.4691806000656, "children": { "RLTrainer._checkpoint": { "total": 0.42435329999989335, "count": 2, "is_parallel": true, "self": 0.42435329999989335 } } }, "_update_policy": { "total": 753.5669778000006, "count": 140, "is_parallel": true, "self": 501.4560800999989, "children": { "TorchPPOOptimizer.update": { "total": 252.11089770000166, "count": 3372, "is_parallel": true, "self": 252.11089770000166 } } } } } } } } }, "TrainerController._save_models": { "total": 0.1655068999998548, "count": 1, "self": 0.006168099999740662, "children": { "RLTrainer._checkpoint": { "total": 0.15933880000011413, "count": 1, "self": 0.15933880000011413 } } } } } } }