poca-SoccerTwos / run_logs /timers.json
dasaprakashk's picture
Initial Commit
c63703c
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.1518149375915527,
"min": 3.1280832290649414,
"max": 3.295764923095703,
"count": 50
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 62532.0078125,
"min": 7487.9755859375,
"max": 108962.4609375,
"count": 50
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 394.38461538461536,
"min": 333.64285714285717,
"max": 999.0,
"count": 50
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 20508.0,
"min": 13232.0,
"max": 27188.0,
"count": 50
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1217.783430049334,
"min": 1198.6947420967692,
"max": 1219.1502834155515,
"count": 49
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 26791.235461085344,
"min": 2399.6648813132306,
"max": 31659.675669988363,
"count": 49
},
"SoccerTwos.Step.mean": {
"value": 499342.0,
"min": 9610.0,
"max": 499342.0,
"count": 50
},
"SoccerTwos.Step.sum": {
"value": 499342.0,
"min": 9610.0,
"max": 499342.0,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.001771173789165914,
"min": -0.0587509423494339,
"max": 0.008092700503766537,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 0.04605051875114441,
"min": -1.3193058967590332,
"max": 0.16994671523571014,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.0030848283786326647,
"min": -0.05528537184000015,
"max": 0.011168592609465122,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -0.080205537378788,
"min": -1.1658110618591309,
"max": 0.23374596238136292,
"count": 50
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 50
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.14319230730716997,
"min": -0.6622699990868568,
"max": 0.38508888747957015,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -3.7229999899864197,
"min": -13.270399957895279,
"max": 6.931599974632263,
"count": 50
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.14319230730716997,
"min": -0.6622699990868568,
"max": 0.38508888747957015,
"count": 50
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -3.7229999899864197,
"min": -13.270399957895279,
"max": 6.931599974632263,
"count": 50
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 50
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 50
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.019773958838777618,
"min": 0.014282445534384654,
"max": 0.021488917719883222,
"count": 22
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.019773958838777618,
"min": 0.014282445534384654,
"max": 0.021488917719883222,
"count": 22
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.0035961836498851576,
"min": 0.00197239763136687,
"max": 0.010879720995823542,
"count": 22
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.0035961836498851576,
"min": 0.00197239763136687,
"max": 0.010879720995823542,
"count": 22
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.0038008132095759114,
"min": 0.0018043413251039705,
"max": 0.011257337064792712,
"count": 22
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.0038008132095759114,
"min": 0.0018043413251039705,
"max": 0.011257337064792712,
"count": 22
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 22
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 22
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 22
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 22
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 22
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 22
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1679237840",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-env-executables/linux/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics --force",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1679238979"
},
"total": 1139.19451972,
"count": 1,
"self": 0.4422080589997677,
"children": {
"run_training.setup": {
"total": 0.10711526600016441,
"count": 1,
"self": 0.10711526600016441
},
"TrainerController.start_learning": {
"total": 1138.645196395,
"count": 1,
"self": 0.9060929790161936,
"children": {
"TrainerController._reset_env": {
"total": 10.464037975999872,
"count": 3,
"self": 10.464037975999872
},
"TrainerController.advance": {
"total": 1127.0194457709838,
"count": 32372,
"self": 1.0659959959739354,
"children": {
"env_step": {
"total": 920.126967747969,
"count": 32372,
"self": 724.232329287993,
"children": {
"SubprocessEnvManager._take_step": {
"total": 195.33478310100145,
"count": 32372,
"self": 6.310763247022123,
"children": {
"TorchPolicy.evaluate": {
"total": 189.02401985397933,
"count": 64076,
"self": 189.02401985397933
}
}
},
"workers": {
"total": 0.5598553589745734,
"count": 32372,
"self": 0.0,
"children": {
"worker_root": {
"total": 1135.4949169180063,
"count": 32372,
"is_parallel": true,
"self": 528.9950504610001,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.006792135999830862,
"count": 2,
"is_parallel": true,
"self": 0.004350616000010632,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0024415199998202297,
"count": 8,
"is_parallel": true,
"self": 0.0024415199998202297
}
}
},
"UnityEnvironment.step": {
"total": 0.06933690400001069,
"count": 1,
"is_parallel": true,
"self": 0.0009805639997466642,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0007959760000630922,
"count": 1,
"is_parallel": true,
"self": 0.0007959760000630922
},
"communicator.exchange": {
"total": 0.06406958000002305,
"count": 1,
"is_parallel": true,
"self": 0.06406958000002305
},
"steps_from_proto": {
"total": 0.0034907840001778823,
"count": 2,
"is_parallel": true,
"self": 0.000761014000090654,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0027297700000872283,
"count": 8,
"is_parallel": true,
"self": 0.0027297700000872283
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 606.4953913580061,
"count": 32371,
"is_parallel": true,
"self": 33.801688653018346,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 25.64243305999662,
"count": 32371,
"is_parallel": true,
"self": 25.64243305999662
},
"communicator.exchange": {
"total": 437.3293248820107,
"count": 32371,
"is_parallel": true,
"self": 437.3293248820107
},
"steps_from_proto": {
"total": 109.72194476298046,
"count": 64742,
"is_parallel": true,
"self": 20.71364357187258,
"children": {
"_process_rank_one_or_two_observation": {
"total": 89.00830119110788,
"count": 258968,
"is_parallel": true,
"self": 89.00830119110788
}
}
}
}
},
"steps_from_proto": {
"total": 0.004475099000046612,
"count": 4,
"is_parallel": true,
"self": 0.000881249999110878,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.003593849000935734,
"count": 16,
"is_parallel": true,
"self": 0.003593849000935734
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 205.8264820270408,
"count": 32372,
"self": 7.714931302054993,
"children": {
"process_trajectory": {
"total": 57.23321689998488,
"count": 32372,
"self": 56.99616937798487,
"children": {
"RLTrainer._checkpoint": {
"total": 0.23704752200001167,
"count": 1,
"self": 0.23704752200001167
}
}
},
"_update_policy": {
"total": 140.87833382500094,
"count": 22,
"self": 87.20122397999853,
"children": {
"TorchPOCAOptimizer.update": {
"total": 53.677109845002406,
"count": 672,
"self": 53.677109845002406
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.0140001904801466e-06,
"count": 1,
"self": 1.0140001904801466e-06
},
"TrainerController._save_models": {
"total": 0.2556186550000348,
"count": 1,
"self": 0.001857178999671305,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2537614760003635,
"count": 1,
"self": 0.2537614760003635
}
}
}
}
}
}
}