poca-SoccerTwos / run_logs /timers.json
JamesEJarvis's picture
Second Push
b1fda6f
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.2650021314620972,
"min": 1.2380719184875488,
"max": 1.468640923500061,
"count": 1000
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 23761.80078125,
"min": 21988.466796875,
"max": 32897.56640625,
"count": 1000
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 67.27397260273973,
"min": 51.583333333333336,
"max": 116.33333333333333,
"count": 1000
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19644.0,
"min": 18228.0,
"max": 21312.0,
"count": 1000
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 2004.5787994439306,
"min": 1954.8947872617402,
"max": 2041.764258181336,
"count": 1000
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 292668.50471881387,
"min": 169423.43903436014,
"max": 383670.03688610066,
"count": 1000
},
"SoccerTwos.Step.mean": {
"value": 84999955.0,
"min": 75009944.0,
"max": 84999955.0,
"count": 1000
},
"SoccerTwos.Step.sum": {
"value": 84999955.0,
"min": 75009944.0,
"max": 84999955.0,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.03205849230289459,
"min": -0.11903506517410278,
"max": 0.059358175843954086,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -4.680540084838867,
"min": -16.39316749572754,
"max": 9.257532119750977,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.033613935112953186,
"min": -0.1195150688290596,
"max": 0.05932600796222687,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -4.907634735107422,
"min": -16.774024963378906,
"max": 9.482162475585938,
"count": 1000
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1000
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.07270958856360553,
"min": -0.43179082651750755,
"max": 0.3916580609736904,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -10.615599930286407,
"min": -59.70240008831024,
"max": 52.31599986553192,
"count": 1000
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.07270958856360553,
"min": -0.43179082651750755,
"max": 0.3916580609736904,
"count": 1000
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -10.615599930286407,
"min": -59.70240008831024,
"max": 52.31599986553192,
"count": 1000
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1000
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1000
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.019802655400902345,
"min": 0.011881238495455667,
"max": 0.027653722364145022,
"count": 485
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.019802655400902345,
"min": 0.011881238495455667,
"max": 0.027653722364145022,
"count": 485
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.09014766265948614,
"min": 0.06759478511909643,
"max": 0.11066784386833509,
"count": 485
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.09014766265948614,
"min": 0.06759478511909643,
"max": 0.11066784386833509,
"count": 485
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.09081148828069369,
"min": 0.06813100427389145,
"max": 0.1114325650036335,
"count": 485
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.09081148828069369,
"min": 0.06813100427389145,
"max": 0.1114325650036335,
"count": 485
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 485
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 485
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 485
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 485
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 485
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 485
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1681230634",
"python_version": "3.9.16 (main, Mar 8 2023, 10:39:24) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\JJarvis-ADM\\.conda\\envs\\rl2\\Scripts\\mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos --no-graphics --resume",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.0.0+cpu",
"numpy_version": "1.21.2",
"end_time_seconds": "1681467910"
},
"total": 237278.1721994,
"count": 1,
"self": 0.49782279998180456,
"children": {
"run_training.setup": {
"total": 0.11812080000000158,
"count": 1,
"self": 0.11812080000000158
},
"TrainerController.start_learning": {
"total": 237277.5562558,
"count": 1,
"self": 12.676609797141282,
"children": {
"TrainerController._reset_env": {
"total": 5.2566997000159486,
"count": 24,
"self": 5.2566997000159486
},
"TrainerController.advance": {
"total": 237259.45996180284,
"count": 680892,
"self": 13.211910992307821,
"children": {
"env_step": {
"total": 31382.051953505576,
"count": 680892,
"self": 29217.357222009934,
"children": {
"SubprocessEnvManager._take_step": {
"total": 2157.2844567964285,
"count": 680892,
"self": 76.64871019214206,
"children": {
"TorchPolicy.evaluate": {
"total": 2080.6357466042864,
"count": 1252470,
"self": 2080.6357466042864
}
}
},
"workers": {
"total": 7.410274699214369,
"count": 680892,
"self": 0.0,
"children": {
"worker_root": {
"total": 237252.86714909453,
"count": 680892,
"is_parallel": true,
"self": 209447.5869146967,
"children": {
"steps_from_proto": {
"total": 0.036516600062578064,
"count": 48,
"is_parallel": true,
"self": 0.007270999944183565,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0292456001183945,
"count": 192,
"is_parallel": true,
"self": 0.0292456001183945
}
}
},
"UnityEnvironment.step": {
"total": 27805.24371779777,
"count": 680892,
"is_parallel": true,
"self": 512.3516042778283,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 527.21155800362,
"count": 680892,
"is_parallel": true,
"self": 527.21155800362
},
"communicator.exchange": {
"total": 25663.8106615048,
"count": 680892,
"is_parallel": true,
"self": 25663.8106615048
},
"steps_from_proto": {
"total": 1101.869894011526,
"count": 1361784,
"is_parallel": true,
"self": 216.25630740287716,
"children": {
"_process_rank_one_or_two_observation": {
"total": 885.6135866086488,
"count": 5447136,
"is_parallel": true,
"self": 885.6135866086488
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 205864.19609730496,
"count": 680892,
"self": 91.93224430960254,
"children": {
"process_trajectory": {
"total": 86885.65257979528,
"count": 680892,
"self": 86882.21135209531,
"children": {
"RLTrainer._checkpoint": {
"total": 3.4412276999578353,
"count": 20,
"self": 3.4412276999578353
}
}
},
"_update_policy": {
"total": 118886.61127320006,
"count": 485,
"self": 1067.9659013991914,
"children": {
"TorchPOCAOptimizer.update": {
"total": 117818.64537180087,
"count": 14550,
"self": 117818.64537180087
}
}
}
}
}
}
},
"trainer_threads": {
"total": 6.00004568696022e-07,
"count": 1,
"self": 6.00004568696022e-07
},
"TrainerController._save_models": {
"total": 0.16298389999428764,
"count": 1,
"self": 0.005163499998161569,
"children": {
"RLTrainer._checkpoint": {
"total": 0.15782039999612607,
"count": 1,
"self": 0.15782039999612607
}
}
}
}
}
}
}