{ "name": "root", "gauges": { "Pyramids.Policy.Entropy.mean": { "value": 0.41989290714263916, "min": 0.4041016399860382, "max": 1.5296342372894287, "count": 40 }, "Pyramids.Policy.Entropy.sum": { "value": 10420.0625, "min": 10157.4990234375, "max": 40137.6015625, "count": 40 }, "Pyramids.Step.mean": { "value": 999891.0, "min": 24960.0, "max": 999891.0, "count": 40 }, "Pyramids.Step.sum": { "value": 999891.0, "min": 24960.0, "max": 999891.0, "count": 40 }, "Pyramids.Policy.ExtrinsicValueEstimate.mean": { "value": 0.576248049736023, "min": -0.06317688524723053, "max": 0.645792543888092, "count": 40 }, "Pyramids.Policy.ExtrinsicValueEstimate.sum": { "value": 133.11329650878906, "min": -12.761730194091797, "max": 152.40704345703125, "count": 40 }, "Pyramids.Policy.RndValueEstimate.mean": { "value": 0.03651592507958412, "min": 0.029619131237268448, "max": 0.8051331639289856, "count": 40 }, "Pyramids.Policy.RndValueEstimate.sum": { "value": 8.435178756713867, "min": 6.871638298034668, "max": 162.63690185546875, "count": 40 }, "Pyramids.Losses.PolicyLoss.mean": { "value": 0.04717810770671349, "min": 0.0452370978859512, "max": 0.05795679721594335, "count": 40 }, "Pyramids.Losses.PolicyLoss.sum": { "value": 0.28306864624028094, "min": 0.20334083630733962, "max": 0.3191777245568422, "count": 40 }, "Pyramids.Losses.ValueLoss.mean": { "value": 0.015732989766143467, "min": 0.0013091506492249512, "max": 0.027568788947202404, "count": 40 }, "Pyramids.Losses.ValueLoss.sum": { "value": 0.09439793859686081, "min": 0.007854903895349707, "max": 0.11027515578880961, "count": 40 }, "Pyramids.Policy.LearningRate.mean": { "value": 3.699848766750002e-06, "min": 3.699848766750002e-06, "max": 0.00029544960151679995, "count": 40 }, "Pyramids.Policy.LearningRate.sum": { "value": 2.2199092600500014e-05, "min": 2.2199092600500014e-05, "max": 0.0016864524378491998, "count": 40 }, "Pyramids.Policy.Epsilon.mean": { "value": 0.10123325, "min": 0.10123325, "max": 0.1984832, "count": 40 }, "Pyramids.Policy.Epsilon.sum": { "value": 0.6073995, "min": 0.6073995, "max": 1.1621508000000003, "count": 40 }, "Pyramids.Policy.Beta.mean": { "value": 0.00013320167500000007, "min": 0.00013320167500000007, "max": 0.00984847168, "count": 40 }, "Pyramids.Policy.Beta.sum": { "value": 0.0007992100500000004, "min": 0.0007992100500000004, "max": 0.05621886492000001, "count": 40 }, "Pyramids.Losses.RNDLoss.mean": { "value": 0.025212058797478676, "min": 0.025212058797478676, "max": 1.175389051437378, "count": 40 }, "Pyramids.Losses.RNDLoss.sum": { "value": 0.15127235651016235, "min": 0.15127235651016235, "max": 4.701556205749512, "count": 40 }, "Pyramids.Environment.EpisodeLength.mean": { "value": 333.56756756756755, "min": 271.8953488372093, "max": 999.0, "count": 40 }, "Pyramids.Environment.EpisodeLength.sum": { "value": 24684.0, "min": 15984.0, "max": 32477.0, "count": 40 }, "Pyramids.Environment.CumulativeReward.mean": { "value": 1.5853243084574067, "min": -1.0000000521540642, "max": 1.7071861869644844, "count": 40 }, "Pyramids.Environment.CumulativeReward.sum": { "value": 117.3139988258481, "min": -26.915001556277275, "max": 148.52519826591015, "count": 40 }, "Pyramids.Policy.ExtrinsicReward.mean": { "value": 1.5853243084574067, "min": -1.0000000521540642, "max": 1.7071861869644844, "count": 40 }, "Pyramids.Policy.ExtrinsicReward.sum": { "value": 117.3139988258481, "min": -26.915001556277275, "max": 148.52519826591015, "count": 40 }, "Pyramids.Policy.RndReward.mean": { "value": 0.08592352152835873, "min": 0.07426628947991279, "max": 23.738284043967724, "count": 40 }, "Pyramids.Policy.RndReward.sum": { "value": 6.358340593098546, "min": 6.358340593098546, "max": 379.8125447034836, "count": 40 }, "Pyramids.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Pyramids.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1676670243", "python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1676672443" }, "total": 2199.86335695, "count": 1, "self": 1.03717770799949, "children": { "run_training.setup": { "total": 0.11201102099994387, "count": 1, "self": 0.11201102099994387 }, "TrainerController.start_learning": { "total": 2198.7141682210004, "count": 1, "self": 1.2552300608485893, "children": { "TrainerController._reset_env": { "total": 7.400739459999841, "count": 1, "self": 7.400739459999841 }, "TrainerController.advance": { "total": 2189.871965095151, "count": 63937, "self": 1.39744750633281, "children": { "env_step": { "total": 1542.4651642069316, "count": 63937, "self": 1425.849401070735, "children": { "SubprocessEnvManager._take_step": { "total": 115.82698808105351, "count": 63937, "self": 4.5958274890786015, "children": { "TorchPolicy.evaluate": { "total": 111.23116059197491, "count": 62556, "self": 35.81660614482007, "children": { "TorchPolicy.sample_actions": { "total": 75.41455444715484, "count": 62556, "self": 75.41455444715484 } } } } }, "workers": { "total": 0.7887750551431054, "count": 63937, "self": 0.0, "children": { "worker_root": { "total": 2193.6409471760117, "count": 63937, "is_parallel": true, "self": 880.8368568800511, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.001856887000030838, "count": 1, "is_parallel": true, "self": 0.0006545509995703469, "children": { "_process_rank_one_or_two_observation": { "total": 0.001202336000460491, "count": 8, "is_parallel": true, "self": 0.001202336000460491 } } }, "UnityEnvironment.step": { "total": 0.06393890299978011, "count": 1, "is_parallel": true, "self": 0.0005791669996142446, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0004989499998373503, "count": 1, "is_parallel": true, "self": 0.0004989499998373503 }, "communicator.exchange": { "total": 0.061080368000148155, "count": 1, "is_parallel": true, "self": 0.061080368000148155 }, "steps_from_proto": { "total": 0.0017804180001803616, "count": 1, "is_parallel": true, "self": 0.0004388479997032846, "children": { "_process_rank_one_or_two_observation": { "total": 0.001341570000477077, "count": 8, "is_parallel": true, "self": 0.001341570000477077 } } } } } } }, "UnityEnvironment.step": { "total": 1312.8040902959606, "count": 63936, "is_parallel": true, "self": 31.625369493765447, "children": { "UnityEnvironment._generate_step_input": { "total": 22.859380549893558, "count": 63936, "is_parallel": true, "self": 22.859380549893558 }, "communicator.exchange": { "total": 1164.8508445391603, "count": 63936, "is_parallel": true, "self": 1164.8508445391603 }, "steps_from_proto": { "total": 93.46849571314124, "count": 63936, "is_parallel": true, "self": 21.65210246877359, "children": { "_process_rank_one_or_two_observation": { "total": 71.81639324436765, "count": 511488, "is_parallel": true, "self": 71.81639324436765 } } } } } } } } } } }, "trainer_advance": { "total": 646.0093533818867, "count": 63937, "self": 2.3960401339513737, "children": { "process_trajectory": { "total": 164.1898366749374, "count": 63937, "self": 163.87566051893646, "children": { "RLTrainer._checkpoint": { "total": 0.31417615600093995, "count": 2, "self": 0.31417615600093995 } } }, "_update_policy": { "total": 479.42347657299797, "count": 234, "self": 221.95328288397423, "children": { "TorchPPOOptimizer.update": { "total": 257.47019368902374, "count": 11511, "self": 257.47019368902374 } } } } } } }, "trainer_threads": { "total": 1.223000253958162e-06, "count": 1, "self": 1.223000253958162e-06 }, "TrainerController._save_models": { "total": 0.18623238200052583, "count": 1, "self": 0.002947506000055, "children": { "RLTrainer._checkpoint": { "total": 0.18328487600047083, "count": 1, "self": 0.18328487600047083 } } } } } } }