poca-SoccerTwos / run_logs /timers.json
Pandita-IA's picture
ADD FILES
ce969cb verified
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.5668424367904663,
"min": 1.5011452436447144,
"max": 3.2958030700683594,
"count": 1500
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 31437.126953125,
"min": 20539.86328125,
"max": 139477.171875,
"count": 1500
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 49.535353535353536,
"min": 37.40769230769231,
"max": 999.0,
"count": 1500
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19616.0,
"min": 13784.0,
"max": 30080.0,
"count": 1500
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1613.1626451644208,
"min": 1182.5759922505104,
"max": 1693.1391499720842,
"count": 1461
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 319406.20374255534,
"min": 2366.6584398141267,
"max": 418503.31246517465,
"count": 1461
},
"SoccerTwos.Step.mean": {
"value": 14999991.0,
"min": 9056.0,
"max": 14999991.0,
"count": 1500
},
"SoccerTwos.Step.sum": {
"value": 14999991.0,
"min": 9056.0,
"max": 14999991.0,
"count": 1500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.042363736778497696,
"min": -0.12342528253793716,
"max": 0.21824456751346588,
"count": 1500
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -8.388019561767578,
"min": -23.450803756713867,
"max": 29.76882553100586,
"count": 1500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.04421196132898331,
"min": -0.12307731062173843,
"max": 0.21696819365024567,
"count": 1500
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -8.753968238830566,
"min": -23.384689331054688,
"max": 30.355152130126953,
"count": 1500
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1500
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1500
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.03473737534850535,
"min": -0.5714285714285714,
"max": 0.5861288127252611,
"count": 1500
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -6.878000319004059,
"min": -59.04639995098114,
"max": 55.0628000497818,
"count": 1500
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.03473737534850535,
"min": -0.5714285714285714,
"max": 0.5861288127252611,
"count": 1500
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -6.878000319004059,
"min": -59.04639995098114,
"max": 55.0628000497818,
"count": 1500
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1500
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1500
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.01692735132625482,
"min": 0.008673173755717774,
"max": 0.02537493456620723,
"count": 724
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.01692735132625482,
"min": 0.008673173755717774,
"max": 0.02537493456620723,
"count": 724
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.1095173495511214,
"min": 2.8146574327555147e-07,
"max": 0.13110793729623157,
"count": 724
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.1095173495511214,
"min": 2.8146574327555147e-07,
"max": 0.13110793729623157,
"count": 724
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.11108667875329653,
"min": 2.4219187485149024e-07,
"max": 0.1342211882273356,
"count": 724
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.11108667875329653,
"min": 2.4219187485149024e-07,
"max": 0.1342211882273356,
"count": 724
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.00019999999999999996,
"min": 0.00019999999999999996,
"max": 0.00019999999999999996,
"count": 724
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.00019999999999999996,
"min": 0.00019999999999999996,
"max": 0.00019999999999999996,
"count": 724
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 724
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 724
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 724
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 724
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1719349536",
"python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
"command_line_arguments": "/home/pandita/anaconda3/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id==SoccerTwosTrial --force --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.3.1+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1719366982"
},
"total": 17446.608223610998,
"count": 1,
"self": 0.21834589999707532,
"children": {
"run_training.setup": {
"total": 0.010502066998014925,
"count": 1,
"self": 0.010502066998014925
},
"TrainerController.start_learning": {
"total": 17446.379375644003,
"count": 1,
"self": 17.780386075039132,
"children": {
"TrainerController._reset_env": {
"total": 2.5676018060075876,
"count": 75,
"self": 2.5676018060075876
},
"TrainerController.advance": {
"total": 17425.867149970956,
"count": 1035080,
"self": 17.467622792311886,
"children": {
"env_step": {
"total": 12158.26244885652,
"count": 1035080,
"self": 8894.618521874247,
"children": {
"SubprocessEnvManager._take_step": {
"total": 3252.2571331651343,
"count": 1035080,
"self": 100.05797721151976,
"children": {
"TorchPolicy.evaluate": {
"total": 3152.1991559536145,
"count": 1889688,
"self": 3152.1991559536145
}
}
},
"workers": {
"total": 11.386793817138823,
"count": 1035080,
"self": 0.0,
"children": {
"worker_root": {
"total": 17420.066335439395,
"count": 1035080,
"is_parallel": true,
"self": 10324.29204544552,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.001568653002323117,
"count": 2,
"is_parallel": true,
"self": 0.0004486799989535939,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0011199730033695232,
"count": 8,
"is_parallel": true,
"self": 0.0011199730033695232
}
}
},
"UnityEnvironment.step": {
"total": 0.015322688002925133,
"count": 1,
"is_parallel": true,
"self": 0.00035246599873062223,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00030690499988850206,
"count": 1,
"is_parallel": true,
"self": 0.00030690499988850206
},
"communicator.exchange": {
"total": 0.013634989001729991,
"count": 1,
"is_parallel": true,
"self": 0.013634989001729991
},
"steps_from_proto": {
"total": 0.0010283280025760178,
"count": 2,
"is_parallel": true,
"self": 0.00023674400654272176,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.000791583996033296,
"count": 8,
"is_parallel": true,
"self": 0.000791583996033296
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 7095.698627933849,
"count": 1035079,
"is_parallel": true,
"self": 361.7603478302808,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 274.0821232143244,
"count": 1035079,
"is_parallel": true,
"self": 274.0821232143244
},
"communicator.exchange": {
"total": 5442.701532716714,
"count": 1035079,
"is_parallel": true,
"self": 5442.701532716714
},
"steps_from_proto": {
"total": 1017.15462417253,
"count": 2070158,
"is_parallel": true,
"self": 200.8899587075939,
"children": {
"_process_rank_one_or_two_observation": {
"total": 816.2646654649361,
"count": 8280632,
"is_parallel": true,
"self": 816.2646654649361
}
}
}
}
},
"steps_from_proto": {
"total": 0.0756620600259339,
"count": 148,
"is_parallel": true,
"self": 0.014940235047106398,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0607218249788275,
"count": 592,
"is_parallel": true,
"self": 0.0607218249788275
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 5250.137078322125,
"count": 1035080,
"self": 137.34018241210288,
"children": {
"process_trajectory": {
"total": 1789.6481766791367,
"count": 1035080,
"self": 1784.776113835127,
"children": {
"RLTrainer._checkpoint": {
"total": 4.872062844009633,
"count": 30,
"self": 4.872062844009633
}
}
},
"_update_policy": {
"total": 3323.148719230885,
"count": 724,
"self": 1161.5716940290258,
"children": {
"TorchPOCAOptimizer.update": {
"total": 2161.577025201859,
"count": 21720,
"self": 2161.577025201859
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.019989425316453e-07,
"count": 1,
"self": 7.019989425316453e-07
},
"TrainerController._save_models": {
"total": 0.16423709000082454,
"count": 1,
"self": 0.0019961860016337596,
"children": {
"RLTrainer._checkpoint": {
"total": 0.16224090399919078,
"count": 1,
"self": 0.16224090399919078
}
}
}
}
}
}
}