poca-SoccerTwos / run_logs /timers.json
slava-medvedev's picture
First Push
e85f9f9
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.681774616241455,
"min": 1.4228068590164185,
"max": 1.681774616241455,
"count": 442
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 37564.1171875,
"min": 19147.625,
"max": 37564.1171875,
"count": 442
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 75.64615384615385,
"min": 49.47,
"max": 120.14634146341463,
"count": 442
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19668.0,
"min": 11464.0,
"max": 21020.0,
"count": 442
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1607.3167432119076,
"min": 1539.2900961020086,
"max": 1642.6004916186455,
"count": 442
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 208951.176617548,
"min": 93582.13269320746,
"max": 321711.5127240893,
"count": 442
},
"SoccerTwos.Step.mean": {
"value": 32379986.0,
"min": 27969991.0,
"max": 32379986.0,
"count": 442
},
"SoccerTwos.Step.sum": {
"value": 32379986.0,
"min": 27969991.0,
"max": 32379986.0,
"count": 442
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.012706947512924671,
"min": -0.10485954582691193,
"max": 0.06124389171600342,
"count": 442
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -1.6519031524658203,
"min": -14.869817733764648,
"max": 9.247827529907227,
"count": 442
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.011308852583169937,
"min": -0.10410598665475845,
"max": 0.06146852299571037,
"count": 442
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -1.4701508283615112,
"min": -15.30374813079834,
"max": 9.281746864318848,
"count": 442
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 442
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 442
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.018182764603541447,
"min": -0.3372515398012081,
"max": 0.30086127884926334,
"count": 442
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -2.363759398460388,
"min": -47.889718651771545,
"max": 39.081418633461,
"count": 442
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.020203075959132268,
"min": -0.3747239440259799,
"max": 0.33429032083480587,
"count": 442
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -2.626399874687195,
"min": -53.21080005168915,
"max": 43.42379981279373,
"count": 442
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 442
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 442
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.016750464675715192,
"min": 0.010129492803631972,
"max": 0.023282602976541965,
"count": 214
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.016750464675715192,
"min": 0.010129492803631972,
"max": 0.023282602976541965,
"count": 214
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.08177757660547892,
"min": 0.058861421048641206,
"max": 0.08222785914937655,
"count": 214
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.08177757660547892,
"min": 0.058861421048641206,
"max": 0.08222785914937655,
"count": 214
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.08331996724009513,
"min": 0.05954822873075803,
"max": 0.08331996724009513,
"count": 214
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.08331996724009513,
"min": 0.05954822873075803,
"max": 0.08331996724009513,
"count": 214
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 214
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 214
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 214
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 214
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 214
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 214
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1698308161",
"python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
"command_line_arguments": "/home/slava/anaconda3/envs/unit6/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./bin/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics --resume",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.1.0+cu121",
"numpy_version": "1.23.3",
"end_time_seconds": "1698316470"
},
"total": 8309.016252023997,
"count": 1,
"self": 0.12568258099781815,
"children": {
"run_training.setup": {
"total": 0.01682956499280408,
"count": 1,
"self": 0.01682956499280408
},
"TrainerController.start_learning": {
"total": 8308.873739878007,
"count": 1,
"self": 6.943632640031865,
"children": {
"TrainerController._reset_env": {
"total": 1.5826032269833377,
"count": 24,
"self": 1.5826032269833377
},
"TrainerController.advance": {
"total": 8300.168301347978,
"count": 302039,
"self": 6.915395208008704,
"children": {
"env_step": {
"total": 6071.931108444012,
"count": 302039,
"self": 4585.344175677834,
"children": {
"SubprocessEnvManager._take_step": {
"total": 1482.3834781259357,
"count": 302039,
"self": 41.37517006495909,
"children": {
"TorchPolicy.evaluate": {
"total": 1441.0083080609766,
"count": 555386,
"self": 1441.0083080609766
}
}
},
"workers": {
"total": 4.203454640242853,
"count": 302038,
"self": 0.0,
"children": {
"worker_root": {
"total": 8297.603483787185,
"count": 302038,
"is_parallel": true,
"self": 4466.0641263351135,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.002016093989368528,
"count": 2,
"is_parallel": true,
"self": 0.00045612099347636104,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001559972995892167,
"count": 8,
"is_parallel": true,
"self": 0.001559972995892167
}
}
},
"UnityEnvironment.step": {
"total": 0.02208193700062111,
"count": 1,
"is_parallel": true,
"self": 0.0004946859990013763,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0003487369976937771,
"count": 1,
"is_parallel": true,
"self": 0.0003487369976937771
},
"communicator.exchange": {
"total": 0.019757002010010183,
"count": 1,
"is_parallel": true,
"self": 0.019757002010010183
},
"steps_from_proto": {
"total": 0.001481511993915774,
"count": 2,
"is_parallel": true,
"self": 0.00030596600845456123,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0011755459854612127,
"count": 8,
"is_parallel": true,
"self": 0.0011755459854612127
}
}
}
}
}
}
},
"steps_from_proto": {
"total": 0.0404042960726656,
"count": 46,
"is_parallel": true,
"self": 0.007957969020935707,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.032446327051729895,
"count": 184,
"is_parallel": true,
"self": 0.032446327051729895
}
}
},
"UnityEnvironment.step": {
"total": 3831.4989531559986,
"count": 302037,
"is_parallel": true,
"self": 200.51537540873687,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 113.30882144028146,
"count": 302037,
"is_parallel": true,
"self": 113.30882144028146
},
"communicator.exchange": {
"total": 2949.3739783105266,
"count": 302037,
"is_parallel": true,
"self": 2949.3739783105266
},
"steps_from_proto": {
"total": 568.3007779964537,
"count": 604074,
"is_parallel": true,
"self": 107.39510311106278,
"children": {
"_process_rank_one_or_two_observation": {
"total": 460.9056748853909,
"count": 2416296,
"is_parallel": true,
"self": 460.9056748853909
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 2221.321797695957,
"count": 302038,
"self": 48.7204944579571,
"children": {
"process_trajectory": {
"total": 755.5529312518192,
"count": 302038,
"self": 753.99295384383,
"children": {
"RLTrainer._checkpoint": {
"total": 1.5599774079892086,
"count": 9,
"self": 1.5599774079892086
}
}
},
"_update_policy": {
"total": 1417.0483719861804,
"count": 214,
"self": 561.4824724310893,
"children": {
"TorchPOCAOptimizer.update": {
"total": 855.5658995550912,
"count": 6420,
"self": 855.5658995550912
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.0010116966441274e-06,
"count": 1,
"self": 1.0010116966441274e-06
},
"TrainerController._save_models": {
"total": 0.17920166200201493,
"count": 1,
"self": 0.0022420489985961467,
"children": {
"RLTrainer._checkpoint": {
"total": 0.17695961300341878,
"count": 1,
"self": 0.17695961300341878
}
}
}
}
}
}
}