poca-SoccerTwos / run_logs /timers.json
loymcg's picture
FirstPush
29c3804 verified
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.5366628170013428,
"min": 1.4786217212677002,
"max": 3.295769453048706,
"count": 2000
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 31913.4140625,
"min": 18865.548828125,
"max": 115717.078125,
"count": 2000
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 56.895348837209305,
"min": 40.78333333333333,
"max": 999.0,
"count": 2000
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19572.0,
"min": 11988.0,
"max": 28848.0,
"count": 2000
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1793.8447339407894,
"min": 1196.454563685818,
"max": 1838.3803215008752,
"count": 1990
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 308541.2942378158,
"min": 2392.909127371636,
"max": 392770.14559695276,
"count": 1990
},
"SoccerTwos.Step.mean": {
"value": 19999980.0,
"min": 9436.0,
"max": 19999980.0,
"count": 2000
},
"SoccerTwos.Step.sum": {
"value": 19999980.0,
"min": 9436.0,
"max": 19999980.0,
"count": 2000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.07302676886320114,
"min": -0.13469351828098297,
"max": 0.18754036724567413,
"count": 2000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -12.487577438354492,
"min": -23.480995178222656,
"max": 30.184131622314453,
"count": 2000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.07323035597801208,
"min": -0.13920488953590393,
"max": 0.1861257255077362,
"count": 2000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -12.522391319274902,
"min": -24.082447052001953,
"max": 29.696781158447266,
"count": 2000
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2000
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2000
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.03342339135052865,
"min": -0.7058823529411765,
"max": 0.5632486456149334,
"count": 2000
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -5.715399920940399,
"min": -76.2203996181488,
"max": 61.5667998790741,
"count": 2000
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.03342339135052865,
"min": -0.7058823529411765,
"max": 0.5632486456149334,
"count": 2000
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -5.715399920940399,
"min": -76.2203996181488,
"max": 61.5667998790741,
"count": 2000
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2000
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2000
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.018319016048068685,
"min": 0.008851306009455584,
"max": 0.02463775666352982,
"count": 967
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.018319016048068685,
"min": 0.008851306009455584,
"max": 0.02463775666352982,
"count": 967
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.09356170917550723,
"min": 0.00018321399256819858,
"max": 0.12386293311913808,
"count": 967
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.09356170917550723,
"min": 0.00018321399256819858,
"max": 0.12386293311913808,
"count": 967
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.09470423782865206,
"min": 0.0001779184106756778,
"max": 0.128805739680926,
"count": 967
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.09470423782865206,
"min": 0.0001779184106756778,
"max": 0.128805739680926,
"count": 967
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 967
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 967
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 967
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 967
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 967
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 967
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1742009731",
"python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
"command_line_arguments": "/home/mcguire/anaconda3/envs/soccer_twos/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id=PocaSoccerTwos --no-graphics --force",
"mlagents_version": "1.2.0.dev0",
"mlagents_envs_version": "1.2.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.6.0+cu124",
"numpy_version": "1.23.5",
"end_time_seconds": "1742037830"
},
"total": 27284.368274757,
"count": 1,
"self": 0.21784216999367345,
"children": {
"run_training.setup": {
"total": 0.009306212000410596,
"count": 1,
"self": 0.009306212000410596
},
"TrainerController.start_learning": {
"total": 27284.141126375005,
"count": 1,
"self": 19.381230467104615,
"children": {
"TrainerController._reset_env": {
"total": 2.298780919011733,
"count": 100,
"self": 2.298780919011733
},
"TrainerController.advance": {
"total": 27262.315184166888,
"count": 1376685,
"self": 19.362979475041357,
"children": {
"env_step": {
"total": 21366.60897016561,
"count": 1376685,
"self": 13875.630254528718,
"children": {
"SubprocessEnvManager._take_step": {
"total": 7478.685192741215,
"count": 1376685,
"self": 119.49047676773534,
"children": {
"TorchPolicy.evaluate": {
"total": 7359.19471597348,
"count": 2513608,
"self": 7359.19471597348
}
}
},
"workers": {
"total": 12.293522895676688,
"count": 1376685,
"self": 0.0,
"children": {
"worker_root": {
"total": 27246.135402047326,
"count": 1376685,
"is_parallel": true,
"self": 15297.75036717147,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0016328840001733624,
"count": 2,
"is_parallel": true,
"self": 0.0004994110013285535,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001133472998844809,
"count": 8,
"is_parallel": true,
"self": 0.001133472998844809
}
}
},
"UnityEnvironment.step": {
"total": 0.01589491499998985,
"count": 1,
"is_parallel": true,
"self": 0.00028864900014013983,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00019859799976984505,
"count": 1,
"is_parallel": true,
"self": 0.00019859799976984505
},
"communicator.exchange": {
"total": 0.014593034000426996,
"count": 1,
"is_parallel": true,
"self": 0.014593034000426996
},
"steps_from_proto": {
"total": 0.0008146339996528695,
"count": 2,
"is_parallel": true,
"self": 0.000175158999809355,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006394749998435145,
"count": 8,
"is_parallel": true,
"self": 0.0006394749998435145
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 11948.300906657863,
"count": 1376684,
"is_parallel": true,
"self": 377.6264062851387,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 221.45370794911378,
"count": 1376684,
"is_parallel": true,
"self": 221.45370794911378
},
"communicator.exchange": {
"total": 10290.391650161542,
"count": 1376684,
"is_parallel": true,
"self": 10290.391650161542
},
"steps_from_proto": {
"total": 1058.8291422620696,
"count": 2753368,
"is_parallel": true,
"self": 222.61271461168417,
"children": {
"_process_rank_one_or_two_observation": {
"total": 836.2164276503854,
"count": 11013472,
"is_parallel": true,
"self": 836.2164276503854
}
}
}
}
},
"steps_from_proto": {
"total": 0.08412821799447556,
"count": 198,
"is_parallel": true,
"self": 0.01780839300590742,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.06631982498856814,
"count": 792,
"is_parallel": true,
"self": 0.06631982498856814
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 5876.343234526238,
"count": 1376685,
"self": 139.25692467576846,
"children": {
"process_trajectory": {
"total": 3319.943228607529,
"count": 1376685,
"self": 3314.3335918695175,
"children": {
"RLTrainer._checkpoint": {
"total": 5.609636738011432,
"count": 40,
"self": 5.609636738011432
}
}
},
"_update_policy": {
"total": 2417.1430812429408,
"count": 967,
"self": 1088.605287230237,
"children": {
"TorchPOCAOptimizer.update": {
"total": 1328.5377940127037,
"count": 29025,
"self": 1328.5377940127037
}
}
}
}
}
}
},
"trainer_threads": {
"total": 4.1900057112798095e-07,
"count": 1,
"self": 4.1900057112798095e-07
},
"TrainerController._save_models": {
"total": 0.145930402999511,
"count": 1,
"self": 0.0008004979972611181,
"children": {
"RLTrainer._checkpoint": {
"total": 0.14512990500224987,
"count": 1,
"self": 0.14512990500224987
}
}
}
}
}
}
}