{ "name": "root", "gauges": { "SoccerTwos.Policy.Entropy.mean": { "value": 1.370168924331665, "min": 1.3621723651885986, "max": 1.6481094360351562, "count": 150 }, "SoccerTwos.Policy.Entropy.sum": { "value": 53184.4765625, "min": 52949.08203125, "max": 68784.53125, "count": 150 }, "SoccerTwos.Environment.EpisodeLength.mean": { "value": 60.7639751552795, "min": 43.44888888888889, "max": 65.27152317880795, "count": 150 }, "SoccerTwos.Environment.EpisodeLength.sum": { "value": 39132.0, "min": 38032.0, "max": 40472.0, "count": 150 }, "SoccerTwos.Self-play.ELO.mean": { "value": 1617.3430029982194, "min": 1520.8554606717662, "max": 1617.9088988365193, "count": 150 }, "SoccerTwos.Self-play.ELO.sum": { "value": 520784.4469654267, "min": 486311.4017050965, "max": 697165.7116705726, "count": 150 }, "SoccerTwos.Step.mean": { "value": 10999966.0, "min": 8019986.0, "max": 10999966.0, "count": 150 }, "SoccerTwos.Step.sum": { "value": 10999966.0, "min": 8019986.0, "max": 10999966.0, "count": 150 }, "SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": { "value": -0.0037202939856797457, "min": -0.0812564343214035, "max": 0.04819022864103317, "count": 150 }, "SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": { "value": -1.197934627532959, "min": -29.823997497558594, "max": 17.92676544189453, "count": 150 }, "SoccerTwos.Policy.ExtrinsicValueEstimate.mean": { "value": -0.002463374752551317, "min": -0.07928887009620667, "max": 0.05857187882065773, "count": 150 }, "SoccerTwos.Policy.ExtrinsicValueEstimate.sum": { "value": -0.7932066917419434, "min": -28.865787506103516, "max": 21.788738250732422, "count": 150 }, "SoccerTwos.Environment.CumulativeReward.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 150 }, "SoccerTwos.Environment.CumulativeReward.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 150 }, "SoccerTwos.Policy.ExtrinsicReward.mean": { "value": 0.09614285734129248, "min": -0.2250201563502467, "max": 0.17519408544545534, "count": 150 }, "SoccerTwos.Policy.ExtrinsicReward.sum": { "value": 30.95800006389618, "min": -87.08280050754547, "max": 65.17219978570938, "count": 150 }, "SoccerTwos.Environment.GroupCumulativeReward.mean": { "value": 0.09614285734129248, "min": -0.2250201563502467, "max": 0.17519408544545534, "count": 150 }, "SoccerTwos.Environment.GroupCumulativeReward.sum": { "value": 30.95800006389618, "min": -87.08280050754547, "max": 65.17219978570938, "count": 150 }, "SoccerTwos.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 150 }, "SoccerTwos.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 150 }, "SoccerTwos.Losses.PolicyLoss.mean": { "value": 0.010988795561327909, "min": 0.007749085542369963, "max": 0.01596735391697924, "count": 73 }, "SoccerTwos.Losses.PolicyLoss.sum": { "value": 0.010988795561327909, "min": 0.007749085542369963, "max": 0.01596735391697924, "count": 73 }, "SoccerTwos.Losses.ValueLoss.mean": { "value": 0.10884295180439948, "min": 0.1005159263809522, "max": 0.12344984561204911, "count": 73 }, "SoccerTwos.Losses.ValueLoss.sum": { "value": 0.10884295180439948, "min": 0.1005159263809522, "max": 0.12344984561204911, "count": 73 }, "SoccerTwos.Losses.BaselineLoss.mean": { "value": 0.10933500727017721, "min": 0.10199711148937543, "max": 0.12813595980405806, "count": 73 }, "SoccerTwos.Losses.BaselineLoss.sum": { "value": 0.10933500727017721, "min": 0.10199711148937543, "max": 0.12813595980405806, "count": 73 }, "SoccerTwos.Policy.LearningRate.mean": { "value": 2.5591816476358955e-07, "min": 2.5591816476358955e-07, "max": 0.00040347880037414557, "count": 73 }, "SoccerTwos.Policy.LearningRate.sum": { "value": 2.5591816476358955e-07, "min": 2.5591816476358955e-07, "max": 0.00040347880037414557, "count": 73 }, "SoccerTwos.Policy.Epsilon.mean": { "value": 0.10001705454545456, "min": 0.10001705454545456, "max": 0.12689858181818184, "count": 73 }, "SoccerTwos.Policy.Epsilon.sum": { "value": 0.10001705454545456, "min": 0.10001705454545456, "max": 0.12689858181818184, "count": 73 }, "SoccerTwos.Policy.Beta.mean": { "value": 1.0851021818181666e-05, "min": 1.0851021818181666e-05, "max": 0.001352239232727273, "count": 73 }, "SoccerTwos.Policy.Beta.sum": { "value": 1.0851021818181666e-05, "min": 1.0851021818181666e-05, "max": 0.001352239232727273, "count": 73 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1680590532", "python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos --run-id=SoccerTwos --no-graphics --resume", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1680608776" }, "total": 18244.700398821, "count": 1, "self": 0.8127974230046675, "children": { "run_training.setup": { "total": 0.1407692479999696, "count": 1, "self": 0.1407692479999696 }, "TrainerController.start_learning": { "total": 18243.74683215, "count": 1, "self": 8.36491601000671, "children": { "TrainerController._reset_env": { "total": 1.6466050499977882, "count": 16, "self": 1.6466050499977882 }, "TrainerController.advance": { "total": 18233.362541368995, "count": 209947, "self": 9.625406053761253, "children": { "env_step": { "total": 6694.494282979973, "count": 209947, "self": 5573.094266053795, "children": { "SubprocessEnvManager._take_step": { "total": 1116.3923109243171, "count": 209947, "self": 50.02055781596414, "children": { "TorchPolicy.evaluate": { "total": 1066.371753108353, "count": 376132, "self": 1066.371753108353 } } }, "workers": { "total": 5.0077060018609245, "count": 209947, "self": 0.0, "children": { "worker_root": { "total": 18217.271145467497, "count": 209947, "is_parallel": true, "self": 13572.523520246705, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.008307132000027195, "count": 2, "is_parallel": true, "self": 0.004914864000284069, "children": { "_process_rank_one_or_two_observation": { "total": 0.003392267999743126, "count": 8, "is_parallel": true, "self": 0.003392267999743126 } } }, "UnityEnvironment.step": { "total": 0.0486355509999612, "count": 1, "is_parallel": true, "self": 0.0013232850003532803, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0008360799999991286, "count": 1, "is_parallel": true, "self": 0.0008360799999991286 }, "communicator.exchange": { "total": 0.042459676999897056, "count": 1, "is_parallel": true, "self": 0.042459676999897056 }, "steps_from_proto": { "total": 0.004016508999711732, "count": 2, "is_parallel": true, "self": 0.0008630439995158667, "children": { "_process_rank_one_or_two_observation": { "total": 0.0031534650001958653, "count": 8, "is_parallel": true, "self": 0.0031534650001958653 } } } } } } }, "steps_from_proto": { "total": 0.04405527500034623, "count": 30, "is_parallel": true, "self": 0.00849365699673399, "children": { "_process_rank_one_or_two_observation": { "total": 0.03556161800361224, "count": 120, "is_parallel": true, "self": 0.03556161800361224 } } }, "UnityEnvironment.step": { "total": 4644.703569945791, "count": 209946, "is_parallel": true, "self": 271.6796566051098, "children": { "UnityEnvironment._generate_step_input": { "total": 160.02877981058214, "count": 209946, "is_parallel": true, "self": 160.02877981058214 }, "communicator.exchange": { "total": 3388.9793343562533, "count": 209946, "is_parallel": true, "self": 3388.9793343562533 }, "steps_from_proto": { "total": 824.0157991738456, "count": 419892, "is_parallel": true, "self": 155.70450403963127, "children": { "_process_rank_one_or_two_observation": { "total": 668.3112951342143, "count": 1679568, "is_parallel": true, "self": 668.3112951342143 } } } } } } } } } } }, "trainer_advance": { "total": 11529.242852335263, "count": 209947, "self": 58.45404036619584, "children": { "process_trajectory": { "total": 1832.045396983074, "count": 209947, "self": 1830.1671662040715, "children": { "RLTrainer._checkpoint": { "total": 1.8782307790024788, "count": 6, "self": 1.8782307790024788 } } }, "_update_policy": { "total": 9638.743414985993, "count": 73, "self": 769.6370660180455, "children": { "TorchPOCAOptimizer.update": { "total": 8869.106348967947, "count": 2190, "self": 8869.106348967947 } } } } } } }, "trainer_threads": { "total": 1.0889998520724475e-06, "count": 1, "self": 1.0889998520724475e-06 }, "TrainerController._save_models": { "total": 0.3727686319980421, "count": 1, "self": 0.0037860929987800773, "children": { "RLTrainer._checkpoint": { "total": 0.36898253899926203, "count": 1, "self": 0.36898253899926203 } } } } } } }