poca-SoccerTwos / run_logs /timers.json
Takano Tsuyoshi
First Push
0d52f99
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 2.540839672088623,
"min": 2.5261971950531006,
"max": 3.2957510948181152,
"count": 202
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 50654.1796875,
"min": 4316.22119140625,
"max": 116934.25,
"count": 202
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 47.26470588235294,
"min": 47.26470588235294,
"max": 999.0,
"count": 202
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19284.0,
"min": 11040.0,
"max": 30336.0,
"count": 202
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1357.263156624345,
"min": 1198.0086179957034,
"max": 1374.4353929496879,
"count": 194
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 276881.68395136634,
"min": 2397.504366660334,
"max": 276881.68395136634,
"count": 194
},
"SoccerTwos.Step.mean": {
"value": 2019970.0,
"min": 9184.0,
"max": 2019970.0,
"count": 202
},
"SoccerTwos.Step.sum": {
"value": 2019970.0,
"min": 9184.0,
"max": 2019970.0,
"count": 202
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.03797806799411774,
"min": -0.05693298578262329,
"max": 0.18078544735908508,
"count": 202
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 7.78550386428833,
"min": -6.814325332641602,
"max": 22.77528953552246,
"count": 202
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.03893809765577316,
"min": -0.06527984887361526,
"max": 0.18322020769119263,
"count": 202
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 7.982309818267822,
"min": -5.5260396003723145,
"max": 22.524194717407227,
"count": 202
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 202
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 202
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.14288000071921,
"min": -0.6619157916621158,
"max": 0.4512695709000463,
"count": 202
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -29.29040014743805,
"min": -29.29040014743805,
"max": 53.3726002573967,
"count": 202
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.14288000071921,
"min": -0.6619157916621158,
"max": 0.4512695709000463,
"count": 202
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -29.29040014743805,
"min": -29.29040014743805,
"max": 53.3726002573967,
"count": 202
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 202
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 202
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.015721297939307988,
"min": 0.011396586409925172,
"max": 0.02239813228273609,
"count": 95
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.015721297939307988,
"min": 0.011396586409925172,
"max": 0.02239813228273609,
"count": 95
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.08864775920907657,
"min": 4.899438614908528e-05,
"max": 0.08864775920907657,
"count": 95
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.08864775920907657,
"min": 4.899438614908528e-05,
"max": 0.08864775920907657,
"count": 95
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.09042609433333079,
"min": 4.416140848964763e-05,
"max": 0.09042609433333079,
"count": 95
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.09042609433333079,
"min": 4.416140848964763e-05,
"max": 0.09042609433333079,
"count": 95
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 95
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 95
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 95
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 95
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 95
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 95
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1697362904",
"python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/linux/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.0.1+cu118",
"numpy_version": "1.23.5",
"end_time_seconds": "1697368339"
},
"total": 5435.169183075001,
"count": 1,
"self": 0.11581885400028114,
"children": {
"run_training.setup": {
"total": 0.07680972499997551,
"count": 1,
"self": 0.07680972499997551
},
"TrainerController.start_learning": {
"total": 5434.976554496,
"count": 1,
"self": 4.479338171033305,
"children": {
"TrainerController._reset_env": {
"total": 8.986132194000106,
"count": 11,
"self": 8.986132194000106
},
"TrainerController.advance": {
"total": 5420.882518026967,
"count": 132853,
"self": 5.111103535927214,
"children": {
"env_step": {
"total": 4428.80152946302,
"count": 132853,
"self": 3427.9943907399193,
"children": {
"SubprocessEnvManager._take_step": {
"total": 998.0725453839362,
"count": 132853,
"self": 32.63041951080186,
"children": {
"TorchPolicy.evaluate": {
"total": 965.4421258731344,
"count": 257582,
"self": 965.4421258731344
}
}
},
"workers": {
"total": 2.7345933391650306,
"count": 132852,
"self": 0.0,
"children": {
"worker_root": {
"total": 5421.161614741967,
"count": 132852,
"is_parallel": true,
"self": 2559.6601144012866,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0065499489999183425,
"count": 2,
"is_parallel": true,
"self": 0.0038127210000311607,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.002737227999887182,
"count": 8,
"is_parallel": true,
"self": 0.002737227999887182
}
}
},
"UnityEnvironment.step": {
"total": 0.10574694399997497,
"count": 1,
"is_parallel": true,
"self": 0.001325990999930582,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.000933486000008088,
"count": 1,
"is_parallel": true,
"self": 0.000933486000008088
},
"communicator.exchange": {
"total": 0.09941921999995884,
"count": 1,
"is_parallel": true,
"self": 0.09941921999995884
},
"steps_from_proto": {
"total": 0.004068247000077463,
"count": 2,
"is_parallel": true,
"self": 0.0006274660001963639,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.003440780999881099,
"count": 8,
"is_parallel": true,
"self": 0.003440780999881099
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 2861.473861513679,
"count": 132851,
"is_parallel": true,
"self": 178.47058673042739,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 114.70862222205017,
"count": 132851,
"is_parallel": true,
"self": 114.70862222205017
},
"communicator.exchange": {
"total": 2022.7566688270404,
"count": 132851,
"is_parallel": true,
"self": 2022.7566688270404
},
"steps_from_proto": {
"total": 545.537983734161,
"count": 265702,
"is_parallel": true,
"self": 92.65374784828668,
"children": {
"_process_rank_one_or_two_observation": {
"total": 452.8842358858743,
"count": 1062808,
"is_parallel": true,
"self": 452.8842358858743
}
}
}
}
},
"steps_from_proto": {
"total": 0.027638827001510435,
"count": 20,
"is_parallel": true,
"self": 0.0053791540053680365,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0222596729961424,
"count": 80,
"is_parallel": true,
"self": 0.0222596729961424
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 986.9698850280195,
"count": 132852,
"self": 37.75450980921562,
"children": {
"process_trajectory": {
"total": 337.9607851288041,
"count": 132852,
"self": 337.00905692780486,
"children": {
"RLTrainer._checkpoint": {
"total": 0.9517282009992414,
"count": 4,
"self": 0.9517282009992414
}
}
},
"_update_policy": {
"total": 611.2545900899997,
"count": 95,
"self": 366.19240322102087,
"children": {
"TorchPOCAOptimizer.update": {
"total": 245.06218686897887,
"count": 2853,
"self": 245.06218686897887
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.8939999790745787e-06,
"count": 1,
"self": 1.8939999790745787e-06
},
"TrainerController._save_models": {
"total": 0.6285642100001496,
"count": 1,
"self": 0.002479512000718387,
"children": {
"RLTrainer._checkpoint": {
"total": 0.6260846979994312,
"count": 1,
"self": 0.6260846979994312
}
}
}
}
}
}
}