poca-SoccerTwos / run_logs /timers.json
dmenini's picture
11M
91399be
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.370168924331665,
"min": 1.3621723651885986,
"max": 1.6481094360351562,
"count": 150
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 53184.4765625,
"min": 52949.08203125,
"max": 68784.53125,
"count": 150
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 60.7639751552795,
"min": 43.44888888888889,
"max": 65.27152317880795,
"count": 150
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 39132.0,
"min": 38032.0,
"max": 40472.0,
"count": 150
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1617.3430029982194,
"min": 1520.8554606717662,
"max": 1617.9088988365193,
"count": 150
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 520784.4469654267,
"min": 486311.4017050965,
"max": 697165.7116705726,
"count": 150
},
"SoccerTwos.Step.mean": {
"value": 10999966.0,
"min": 8019986.0,
"max": 10999966.0,
"count": 150
},
"SoccerTwos.Step.sum": {
"value": 10999966.0,
"min": 8019986.0,
"max": 10999966.0,
"count": 150
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.0037202939856797457,
"min": -0.0812564343214035,
"max": 0.04819022864103317,
"count": 150
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -1.197934627532959,
"min": -29.823997497558594,
"max": 17.92676544189453,
"count": 150
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.002463374752551317,
"min": -0.07928887009620667,
"max": 0.05857187882065773,
"count": 150
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -0.7932066917419434,
"min": -28.865787506103516,
"max": 21.788738250732422,
"count": 150
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 150
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 150
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.09614285734129248,
"min": -0.2250201563502467,
"max": 0.17519408544545534,
"count": 150
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 30.95800006389618,
"min": -87.08280050754547,
"max": 65.17219978570938,
"count": 150
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.09614285734129248,
"min": -0.2250201563502467,
"max": 0.17519408544545534,
"count": 150
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 30.95800006389618,
"min": -87.08280050754547,
"max": 65.17219978570938,
"count": 150
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 150
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 150
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.010988795561327909,
"min": 0.007749085542369963,
"max": 0.01596735391697924,
"count": 73
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.010988795561327909,
"min": 0.007749085542369963,
"max": 0.01596735391697924,
"count": 73
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.10884295180439948,
"min": 0.1005159263809522,
"max": 0.12344984561204911,
"count": 73
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.10884295180439948,
"min": 0.1005159263809522,
"max": 0.12344984561204911,
"count": 73
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.10933500727017721,
"min": 0.10199711148937543,
"max": 0.12813595980405806,
"count": 73
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.10933500727017721,
"min": 0.10199711148937543,
"max": 0.12813595980405806,
"count": 73
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 2.5591816476358955e-07,
"min": 2.5591816476358955e-07,
"max": 0.00040347880037414557,
"count": 73
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 2.5591816476358955e-07,
"min": 2.5591816476358955e-07,
"max": 0.00040347880037414557,
"count": 73
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.10001705454545456,
"min": 0.10001705454545456,
"max": 0.12689858181818184,
"count": 73
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.10001705454545456,
"min": 0.10001705454545456,
"max": 0.12689858181818184,
"count": 73
},
"SoccerTwos.Policy.Beta.mean": {
"value": 1.0851021818181666e-05,
"min": 1.0851021818181666e-05,
"max": 0.001352239232727273,
"count": 73
},
"SoccerTwos.Policy.Beta.sum": {
"value": 1.0851021818181666e-05,
"min": 1.0851021818181666e-05,
"max": 0.001352239232727273,
"count": 73
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1680590532",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos --run-id=SoccerTwos --no-graphics --resume",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1680608776"
},
"total": 18244.700398821,
"count": 1,
"self": 0.8127974230046675,
"children": {
"run_training.setup": {
"total": 0.1407692479999696,
"count": 1,
"self": 0.1407692479999696
},
"TrainerController.start_learning": {
"total": 18243.74683215,
"count": 1,
"self": 8.36491601000671,
"children": {
"TrainerController._reset_env": {
"total": 1.6466050499977882,
"count": 16,
"self": 1.6466050499977882
},
"TrainerController.advance": {
"total": 18233.362541368995,
"count": 209947,
"self": 9.625406053761253,
"children": {
"env_step": {
"total": 6694.494282979973,
"count": 209947,
"self": 5573.094266053795,
"children": {
"SubprocessEnvManager._take_step": {
"total": 1116.3923109243171,
"count": 209947,
"self": 50.02055781596414,
"children": {
"TorchPolicy.evaluate": {
"total": 1066.371753108353,
"count": 376132,
"self": 1066.371753108353
}
}
},
"workers": {
"total": 5.0077060018609245,
"count": 209947,
"self": 0.0,
"children": {
"worker_root": {
"total": 18217.271145467497,
"count": 209947,
"is_parallel": true,
"self": 13572.523520246705,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.008307132000027195,
"count": 2,
"is_parallel": true,
"self": 0.004914864000284069,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.003392267999743126,
"count": 8,
"is_parallel": true,
"self": 0.003392267999743126
}
}
},
"UnityEnvironment.step": {
"total": 0.0486355509999612,
"count": 1,
"is_parallel": true,
"self": 0.0013232850003532803,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0008360799999991286,
"count": 1,
"is_parallel": true,
"self": 0.0008360799999991286
},
"communicator.exchange": {
"total": 0.042459676999897056,
"count": 1,
"is_parallel": true,
"self": 0.042459676999897056
},
"steps_from_proto": {
"total": 0.004016508999711732,
"count": 2,
"is_parallel": true,
"self": 0.0008630439995158667,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0031534650001958653,
"count": 8,
"is_parallel": true,
"self": 0.0031534650001958653
}
}
}
}
}
}
},
"steps_from_proto": {
"total": 0.04405527500034623,
"count": 30,
"is_parallel": true,
"self": 0.00849365699673399,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.03556161800361224,
"count": 120,
"is_parallel": true,
"self": 0.03556161800361224
}
}
},
"UnityEnvironment.step": {
"total": 4644.703569945791,
"count": 209946,
"is_parallel": true,
"self": 271.6796566051098,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 160.02877981058214,
"count": 209946,
"is_parallel": true,
"self": 160.02877981058214
},
"communicator.exchange": {
"total": 3388.9793343562533,
"count": 209946,
"is_parallel": true,
"self": 3388.9793343562533
},
"steps_from_proto": {
"total": 824.0157991738456,
"count": 419892,
"is_parallel": true,
"self": 155.70450403963127,
"children": {
"_process_rank_one_or_two_observation": {
"total": 668.3112951342143,
"count": 1679568,
"is_parallel": true,
"self": 668.3112951342143
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 11529.242852335263,
"count": 209947,
"self": 58.45404036619584,
"children": {
"process_trajectory": {
"total": 1832.045396983074,
"count": 209947,
"self": 1830.1671662040715,
"children": {
"RLTrainer._checkpoint": {
"total": 1.8782307790024788,
"count": 6,
"self": 1.8782307790024788
}
}
},
"_update_policy": {
"total": 9638.743414985993,
"count": 73,
"self": 769.6370660180455,
"children": {
"TorchPOCAOptimizer.update": {
"total": 8869.106348967947,
"count": 2190,
"self": 8869.106348967947
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.0889998520724475e-06,
"count": 1,
"self": 1.0889998520724475e-06
},
"TrainerController._save_models": {
"total": 0.3727686319980421,
"count": 1,
"self": 0.0037860929987800773,
"children": {
"RLTrainer._checkpoint": {
"total": 0.36898253899926203,
"count": 1,
"self": 0.36898253899926203
}
}
}
}
}
}
}