ppo-Huggy / run_logs /timers.json
austinmw's picture
Huggy
cc7a8a2
raw
history blame
No virus
17.9 kB
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4011471271514893,
"min": 1.4011471271514893,
"max": 1.4276500940322876,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 69614.59375,
"min": 68130.1875,
"max": 77622.953125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 113.73165137614679,
"min": 88.13523131672598,
"max": 396.515873015873,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49587.0,
"min": 49119.0,
"max": 49961.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999979.0,
"min": 49728.0,
"max": 1999979.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999979.0,
"min": 49728.0,
"max": 1999979.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4358789920806885,
"min": 0.23684076964855194,
"max": 2.4650659561157227,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1062.043212890625,
"min": 29.6050968170166,
"max": 1317.886474609375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.67236511778394,
"min": 1.8765153735876083,
"max": 4.00708345127405,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1601.151191353798,
"min": 234.56442169845104,
"max": 2126.467792212963,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.67236511778394,
"min": 1.8765153735876083,
"max": 4.00708345127405,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1601.151191353798,
"min": 234.56442169845104,
"max": 2126.467792212963,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.016960462129173317,
"min": 0.013480592398506512,
"max": 0.02066681460128166,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.05088138638751995,
"min": 0.026961184797013024,
"max": 0.05912554970418569,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04851337762342559,
"min": 0.021118568597982328,
"max": 0.055781113294263676,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.14554013287027678,
"min": 0.042237137195964655,
"max": 0.16476590347786743,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.1944489352166685e-06,
"min": 3.1944489352166685e-06,
"max": 0.00029532540155820003,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 9.583346805650005e-06,
"min": 9.583346805650005e-06,
"max": 0.0008440696686434499,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10106478333333334,
"min": 0.10106478333333334,
"max": 0.19844179999999997,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30319435,
"min": 0.2072832,
"max": 0.5813565500000002,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.313268833333336e-05,
"min": 6.313268833333336e-05,
"max": 0.004922245820000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00018939806500000008,
"min": 0.00018939806500000008,
"max": 0.014069691845,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1675626158",
"python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1675628510"
},
"total": 2352.215073304,
"count": 1,
"self": 0.3945956299999125,
"children": {
"run_training.setup": {
"total": 0.11927082700003666,
"count": 1,
"self": 0.11927082700003666
},
"TrainerController.start_learning": {
"total": 2351.701206847,
"count": 1,
"self": 3.9994322340671715,
"children": {
"TrainerController._reset_env": {
"total": 11.731501403000038,
"count": 1,
"self": 11.731501403000038
},
"TrainerController.advance": {
"total": 2335.8575224479323,
"count": 232281,
"self": 4.468318625960819,
"children": {
"env_step": {
"total": 1813.0852703979874,
"count": 232281,
"self": 1516.293821270835,
"children": {
"SubprocessEnvManager._take_step": {
"total": 293.9966383310359,
"count": 232281,
"self": 15.433152320977626,
"children": {
"TorchPolicy.evaluate": {
"total": 278.5634860100583,
"count": 222966,
"self": 68.50035942606013,
"children": {
"TorchPolicy.sample_actions": {
"total": 210.06312658399816,
"count": 222966,
"self": 210.06312658399816
}
}
}
}
},
"workers": {
"total": 2.794810796116394,
"count": 232281,
"self": 0.0,
"children": {
"worker_root": {
"total": 2343.8872214030002,
"count": 232281,
"is_parallel": true,
"self": 1109.0145026210557,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0060064269999884345,
"count": 1,
"is_parallel": true,
"self": 0.00035450000001446824,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.005651926999973966,
"count": 2,
"is_parallel": true,
"self": 0.005651926999973966
}
}
},
"UnityEnvironment.step": {
"total": 0.03500783499998761,
"count": 1,
"is_parallel": true,
"self": 0.00030714399991893515,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00019252200002028985,
"count": 1,
"is_parallel": true,
"self": 0.00019252200002028985
},
"communicator.exchange": {
"total": 0.03378682700002855,
"count": 1,
"is_parallel": true,
"self": 0.03378682700002855
},
"steps_from_proto": {
"total": 0.0007213420000198312,
"count": 1,
"is_parallel": true,
"self": 0.0002443780000476181,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00047696399997221306,
"count": 2,
"is_parallel": true,
"self": 0.00047696399997221306
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1234.8727187819445,
"count": 232280,
"is_parallel": true,
"self": 34.02525633087316,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 73.29200641698117,
"count": 232280,
"is_parallel": true,
"self": 73.29200641698117
},
"communicator.exchange": {
"total": 1038.0275683280288,
"count": 232280,
"is_parallel": true,
"self": 1038.0275683280288
},
"steps_from_proto": {
"total": 89.52788770606134,
"count": 232280,
"is_parallel": true,
"self": 36.60839902010326,
"children": {
"_process_rank_one_or_two_observation": {
"total": 52.91948868595807,
"count": 464560,
"is_parallel": true,
"self": 52.91948868595807
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 518.303933423984,
"count": 232281,
"self": 7.016489791015147,
"children": {
"process_trajectory": {
"total": 160.69855198796978,
"count": 232281,
"self": 159.34630151896937,
"children": {
"RLTrainer._checkpoint": {
"total": 1.3522504690004098,
"count": 10,
"self": 1.3522504690004098
}
}
},
"_update_policy": {
"total": 350.588891644999,
"count": 97,
"self": 294.3342653429894,
"children": {
"TorchPPOOptimizer.update": {
"total": 56.254626302009626,
"count": 2910,
"self": 56.254626302009626
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.370002433366608e-07,
"count": 1,
"self": 8.370002433366608e-07
},
"TrainerController._save_models": {
"total": 0.11274992500011649,
"count": 1,
"self": 0.002250617000299826,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11049930799981666,
"count": 1,
"self": 0.11049930799981666
}
}
}
}
}
}
}