ppo-Huggy / run_logs /timers.json
kuanyk's picture
Huggy
20843f6
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4053642749786377,
"min": 1.4053642749786377,
"max": 1.4283690452575684,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70619.5546875,
"min": 68708.5078125,
"max": 77826.578125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 102.75933609958506,
"min": 90.74128440366972,
"max": 408.6585365853659,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49530.0,
"min": 49252.0,
"max": 50265.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999926.0,
"min": 49991.0,
"max": 1999926.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999926.0,
"min": 49991.0,
"max": 1999926.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.308931589126587,
"min": 0.17180845141410828,
"max": 2.3894405364990234,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1112.905029296875,
"min": 20.960630416870117,
"max": 1291.5745849609375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.5280743802731465,
"min": 1.7704829380649034,
"max": 3.9432901641959517,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1700.5318512916565,
"min": 215.99891844391823,
"max": 2069.1333928108215,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.5280743802731465,
"min": 1.7704829380649034,
"max": 3.9432901641959517,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1700.5318512916565,
"min": 215.99891844391823,
"max": 2069.1333928108215,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.020181093417340888,
"min": 0.014047633033442253,
"max": 0.020181093417340888,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.040362186834681776,
"min": 0.028095266066884506,
"max": 0.05792311582499679,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04629396237432957,
"min": 0.02057565807675322,
"max": 0.05518561191856861,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.09258792474865914,
"min": 0.04115131615350644,
"max": 0.15849718650182087,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 4.410023530025002e-06,
"min": 4.410023530025002e-06,
"max": 0.00029532360155879996,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 8.820047060050003e-06,
"min": 8.820047060050003e-06,
"max": 0.00084392896869035,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10146997500000003,
"min": 0.10146997500000003,
"max": 0.19844119999999998,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20293995000000006,
"min": 0.20293995000000006,
"max": 0.5813096500000001,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 8.335175250000009e-05,
"min": 8.335175250000009e-05,
"max": 0.004922215880000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00016670350500000018,
"min": 0.00016670350500000018,
"max": 0.014067351535000001,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1671596375",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1671598755"
},
"total": 2380.020735127,
"count": 1,
"self": 0.3864160270004504,
"children": {
"run_training.setup": {
"total": 0.12155529699998624,
"count": 1,
"self": 0.12155529699998624
},
"TrainerController.start_learning": {
"total": 2379.512763803,
"count": 1,
"self": 4.1540167521193325,
"children": {
"TrainerController._reset_env": {
"total": 7.421829070000058,
"count": 1,
"self": 7.421829070000058
},
"TrainerController.advance": {
"total": 2367.816146736881,
"count": 231487,
"self": 4.37723948294115,
"children": {
"env_step": {
"total": 1885.1896391159698,
"count": 231487,
"self": 1586.6147926038934,
"children": {
"SubprocessEnvManager._take_step": {
"total": 295.7757880989868,
"count": 231487,
"self": 15.430778770924235,
"children": {
"TorchPolicy.evaluate": {
"total": 280.3450093280626,
"count": 222974,
"self": 70.26811284696907,
"children": {
"TorchPolicy.sample_actions": {
"total": 210.0768964810935,
"count": 222974,
"self": 210.0768964810935
}
}
}
}
},
"workers": {
"total": 2.799058413089597,
"count": 231487,
"self": 0.0,
"children": {
"worker_root": {
"total": 2371.2141216641044,
"count": 231487,
"is_parallel": true,
"self": 1066.79319089515,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0022178550000262476,
"count": 1,
"is_parallel": true,
"self": 0.0003520569999864165,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001865798000039831,
"count": 2,
"is_parallel": true,
"self": 0.001865798000039831
}
}
},
"UnityEnvironment.step": {
"total": 0.02920985200000814,
"count": 1,
"is_parallel": true,
"self": 0.0002830180000046312,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00022277100003975647,
"count": 1,
"is_parallel": true,
"self": 0.00022277100003975647
},
"communicator.exchange": {
"total": 0.027974368999934995,
"count": 1,
"is_parallel": true,
"self": 0.027974368999934995
},
"steps_from_proto": {
"total": 0.0007296940000287577,
"count": 1,
"is_parallel": true,
"self": 0.0002466990000584701,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0004829949999702876,
"count": 2,
"is_parallel": true,
"self": 0.0004829949999702876
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1304.4209307689543,
"count": 231486,
"is_parallel": true,
"self": 36.01791338009616,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 86.22300103295095,
"count": 231486,
"is_parallel": true,
"self": 86.22300103295095
},
"communicator.exchange": {
"total": 1081.8014842249627,
"count": 231486,
"is_parallel": true,
"self": 1081.8014842249627
},
"steps_from_proto": {
"total": 100.37853213094456,
"count": 231486,
"is_parallel": true,
"self": 43.85555979300466,
"children": {
"_process_rank_one_or_two_observation": {
"total": 56.522972337939905,
"count": 462972,
"is_parallel": true,
"self": 56.522972337939905
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 478.2492681379698,
"count": 231487,
"self": 6.2580538959789465,
"children": {
"process_trajectory": {
"total": 156.04441520699527,
"count": 231487,
"self": 154.81839109699536,
"children": {
"RLTrainer._checkpoint": {
"total": 1.226024109999912,
"count": 10,
"self": 1.226024109999912
}
}
},
"_update_policy": {
"total": 315.9467990349956,
"count": 96,
"self": 262.8362636079902,
"children": {
"TorchPPOOptimizer.update": {
"total": 53.1105354270054,
"count": 2880,
"self": 53.1105354270054
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.869995665212628e-07,
"count": 1,
"self": 9.869995665212628e-07
},
"TrainerController._save_models": {
"total": 0.12077025700000377,
"count": 1,
"self": 0.00200682600006985,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11876343099993392,
"count": 1,
"self": 0.11876343099993392
}
}
}
}
}
}
}