{ "name": "root", "gauges": { "Pyramids.Policy.Entropy.mean": { "value": 0.36727404594421387, "min": 0.36727404594421387, "max": 1.4422281980514526, "count": 33 }, "Pyramids.Policy.Entropy.sum": { "value": 11065.232421875, "min": 11065.232421875, "max": 43751.43359375, "count": 33 }, "Pyramids.Step.mean": { "value": 989981.0, "min": 29874.0, "max": 989981.0, "count": 33 }, "Pyramids.Step.sum": { "value": 989981.0, "min": 29874.0, "max": 989981.0, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.mean": { "value": 0.5059118866920471, "min": -0.2092449963092804, "max": 0.5977255702018738, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.sum": { "value": 136.09030151367188, "min": -49.591064453125, "max": 163.7768096923828, "count": 33 }, "Pyramids.Policy.RndValueEstimate.mean": { "value": 0.02397209219634533, "min": -0.036732375621795654, "max": 0.47359856963157654, "count": 33 }, "Pyramids.Policy.RndValueEstimate.sum": { "value": 6.448493003845215, "min": -9.844276428222656, "max": 112.24285888671875, "count": 33 }, "Pyramids.Losses.PolicyLoss.mean": { "value": 0.06610151844922386, "min": 0.06570405096101153, "max": 0.07302507424120368, "count": 33 }, "Pyramids.Losses.PolicyLoss.sum": { "value": 0.9254212582891341, "min": 0.5842005939296294, "max": 1.0914733522513416, "count": 33 }, "Pyramids.Losses.ValueLoss.mean": { "value": 0.01397524023784735, "min": 0.00043636349157859613, "max": 0.015998304096380405, "count": 33 }, "Pyramids.Losses.ValueLoss.sum": { "value": 0.19565336332986288, "min": 0.004799998407364558, "max": 0.2276985700397442, "count": 33 }, "Pyramids.Policy.LearningRate.mean": { "value": 7.667990301178571e-06, "min": 7.667990301178571e-06, "max": 0.000294848776717075, "count": 33 }, "Pyramids.Policy.LearningRate.sum": { "value": 0.0001073518642165, "min": 0.0001073518642165, "max": 0.0035071664309446, "count": 33 }, "Pyramids.Policy.Epsilon.mean": { "value": 0.1025559642857143, "min": 0.1025559642857143, "max": 0.198282925, "count": 33 }, "Pyramids.Policy.Epsilon.sum": { "value": 1.4357835, "min": 1.4357835, "max": 2.5690554000000003, "count": 33 }, "Pyramids.Policy.Beta.mean": { "value": 0.00026534083214285715, "min": 0.00026534083214285715, "max": 0.0098284642075, "count": 33 }, "Pyramids.Policy.Beta.sum": { "value": 0.0037147716500000004, "min": 0.0037147716500000004, "max": 0.11692863446000003, "count": 33 }, "Pyramids.Losses.RNDLoss.mean": { "value": 0.009687610901892185, "min": 0.009687610901892185, "max": 0.48567667603492737, "count": 33 }, "Pyramids.Losses.RNDLoss.sum": { "value": 0.13562655448913574, "min": 0.13562655448913574, "max": 3.885413408279419, "count": 33 }, "Pyramids.Environment.EpisodeLength.mean": { "value": 388.75, "min": 337.16483516483515, "max": 999.0, "count": 33 }, "Pyramids.Environment.EpisodeLength.sum": { "value": 31100.0, "min": 16417.0, "max": 33774.0, "count": 33 }, "Pyramids.Environment.CumulativeReward.mean": { "value": 1.4572506055613108, "min": -0.9999500522390008, "max": 1.6408461349514813, "count": 33 }, "Pyramids.Environment.CumulativeReward.sum": { "value": 115.12279783934355, "min": -31.998401671648026, "max": 149.3169982805848, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.mean": { "value": 1.4572506055613108, "min": -0.9999500522390008, "max": 1.6408461349514813, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.sum": { "value": 115.12279783934355, "min": -31.998401671648026, "max": 149.3169982805848, "count": 33 }, "Pyramids.Policy.RndReward.mean": { "value": 0.03960668348819253, "min": 0.03605160051254485, "max": 9.885111937628073, "count": 33 }, "Pyramids.Policy.RndReward.sum": { "value": 3.12892799556721, "min": 3.041725446411874, "max": 168.04690293967724, "count": 33 }, "Pyramids.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 }, "Pyramids.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1676890207", "python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1676892462" }, "total": 2255.084519381, "count": 1, "self": 0.5261562840000806, "children": { "run_training.setup": { "total": 0.10941882199995234, "count": 1, "self": 0.10941882199995234 }, "TrainerController.start_learning": { "total": 2254.448944275, "count": 1, "self": 1.3917152860672104, "children": { "TrainerController._reset_env": { "total": 6.157329474000107, "count": 1, "self": 6.157329474000107 }, "TrainerController.advance": { "total": 2246.8122810399323, "count": 63851, "self": 1.4199081048768676, "children": { "env_step": { "total": 1506.7423178100557, "count": 63851, "self": 1393.118303322066, "children": { "SubprocessEnvManager._take_step": { "total": 112.78462862304013, "count": 63851, "self": 4.576450978034018, "children": { "TorchPolicy.evaluate": { "total": 108.20817764500612, "count": 62561, "self": 36.22303356600992, "children": { "TorchPolicy.sample_actions": { "total": 71.9851440789962, "count": 62561, "self": 71.9851440789962 } } } } }, "workers": { "total": 0.8393858649496906, "count": 63851, "self": 0.0, "children": { "worker_root": { "total": 2249.568864687006, "count": 63851, "is_parallel": true, "self": 970.6626404360343, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0018900689999554743, "count": 1, "is_parallel": true, "self": 0.0007378769998922508, "children": { "_process_rank_one_or_two_observation": { "total": 0.0011521920000632235, "count": 8, "is_parallel": true, "self": 0.0011521920000632235 } } }, "UnityEnvironment.step": { "total": 0.04704036499992981, "count": 1, "is_parallel": true, "self": 0.0005296289996294945, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00045404600018628116, "count": 1, "is_parallel": true, "self": 0.00045404600018628116 }, "communicator.exchange": { "total": 0.04446976400004132, "count": 1, "is_parallel": true, "self": 0.04446976400004132 }, "steps_from_proto": { "total": 0.0015869260000727081, "count": 1, "is_parallel": true, "self": 0.0004164200001923746, "children": { "_process_rank_one_or_two_observation": { "total": 0.0011705059998803335, "count": 8, "is_parallel": true, "self": 0.0011705059998803335 } } } } } } }, "UnityEnvironment.step": { "total": 1278.906224250972, "count": 63850, "is_parallel": true, "self": 31.36857501094846, "children": { "UnityEnvironment._generate_step_input": { "total": 22.65845254899341, "count": 63850, "is_parallel": true, "self": 22.65845254899341 }, "communicator.exchange": { "total": 1131.8636955020154, "count": 63850, "is_parallel": true, "self": 1131.8636955020154 }, "steps_from_proto": { "total": 93.01550118901469, "count": 63850, "is_parallel": true, "self": 21.69971797011317, "children": { "_process_rank_one_or_two_observation": { "total": 71.31578321890152, "count": 510800, "is_parallel": true, "self": 71.31578321890152 } } } } } } } } } } }, "trainer_advance": { "total": 738.6500551249994, "count": 63851, "self": 2.5703530890734783, "children": { "process_trajectory": { "total": 162.1225703039322, "count": 63851, "self": 161.9268109859322, "children": { "RLTrainer._checkpoint": { "total": 0.19575931800000035, "count": 2, "self": 0.19575931800000035 } } }, "_update_policy": { "total": 573.9571317319937, "count": 452, "self": 220.78286451500844, "children": { "TorchPPOOptimizer.update": { "total": 353.1742672169853, "count": 22821, "self": 353.1742672169853 } } } } } } }, "trainer_threads": { "total": 8.490001164318528e-07, "count": 1, "self": 8.490001164318528e-07 }, "TrainerController._save_models": { "total": 0.08761762600033762, "count": 1, "self": 0.001369582000279479, "children": { "RLTrainer._checkpoint": { "total": 0.08624804400005814, "count": 1, "self": 0.08624804400005814 } } } } } } }