poca-SoccerTwos / run_logs /timers.json
gmojko's picture
First Push`
775fd9e
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 2.9924488067626953,
"min": 2.958897590637207,
"max": 3.1828112602233887,
"count": 300
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 77660.03125,
"min": 21516.66015625,
"max": 137188.0625,
"count": 300
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 475.54545454545456,
"min": 334.06666666666666,
"max": 999.0,
"count": 300
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 20924.0,
"min": 4976.0,
"max": 30936.0,
"count": 300
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1241.8774269485682,
"min": 1192.8382636217825,
"max": 1247.5560014737573,
"count": 282
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 22353.793685074226,
"min": 2387.111447047758,
"max": 34624.33795049888,
"count": 282
},
"SoccerTwos.Step.mean": {
"value": 2999666.0,
"min": 9484.0,
"max": 2999666.0,
"count": 300
},
"SoccerTwos.Step.sum": {
"value": 2999666.0,
"min": 9484.0,
"max": 2999666.0,
"count": 300
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.016900047659873962,
"min": -0.019958587363362312,
"max": -0.001513068564236164,
"count": 300
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -0.3549010157585144,
"min": -0.3693455159664154,
"max": -0.01775738224387169,
"count": 300
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.01757766678929329,
"min": -0.019750839099287987,
"max": -0.0010202398989349604,
"count": 300
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -0.3691309988498688,
"min": -0.3691309988498688,
"max": -0.012242878787219524,
"count": 300
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 300
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 300
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.11979047457377116,
"min": -0.682741176117869,
"max": 0.338373335202535,
"count": 300
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -2.5155999660491943,
"min": -13.14359998703003,
"max": 5.208000034093857,
"count": 300
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.11979047457377116,
"min": -0.682741176117869,
"max": 0.338373335202535,
"count": 300
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -2.5155999660491943,
"min": -13.14359998703003,
"max": 5.208000034093857,
"count": 300
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 300
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 300
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.008282075870859747,
"min": 0.005789831588238788,
"max": 0.010839468071208103,
"count": 36
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.008282075870859747,
"min": 0.005789831588238788,
"max": 0.010839468071208103,
"count": 36
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.006328290343905489,
"min": 0.001657259821270903,
"max": 0.009326495820035537,
"count": 36
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.006328290343905489,
"min": 0.001657259821270903,
"max": 0.009326495820035537,
"count": 36
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.006342333027472099,
"min": 0.0016584567182386914,
"max": 0.009329031966626645,
"count": 36
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.006342333027472099,
"min": 0.0016584567182386914,
"max": 0.009329031966626645,
"count": 36
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 7.372099385666599e-06,
"min": 7.372099385666599e-06,
"max": 0.0011668776027601998,
"count": 36
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 7.372099385666599e-06,
"min": 7.372099385666599e-06,
"max": 0.0011668776027601998,
"count": 36
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.10061433333333329,
"min": 0.10061433333333329,
"max": 0.19723980000000008,
"count": 36
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.10061433333333329,
"min": 0.10061433333333329,
"max": 0.19723980000000008,
"count": 36
},
"SoccerTwos.Policy.Beta.mean": {
"value": 4.065523333333305e-05,
"min": 4.065523333333305e-05,
"max": 0.004862266020000001,
"count": 36
},
"SoccerTwos.Policy.Beta.sum": {
"value": 4.065523333333305e-05,
"min": 4.065523333333305e-05,
"max": 0.004862266020000001,
"count": 36
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1677574592",
"python_version": "3.9.16 (main, Jan 11 2023, 16:16:36) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\gmojkowski\\.conda\\envs\\rl\\Scripts\\mlagents-learn ml-agents/config/poca/SoccerTwos.yaml --env SoccerTwos/SoccerTwos.exe --run-id=resnet_base_v4 --no-graphics --initialize-from=resnet_base_v3",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.13.1+cpu",
"numpy_version": "1.21.2",
"end_time_seconds": "1677618311"
},
"total": 43718.9919943,
"count": 1,
"self": 0.4404387999966275,
"children": {
"run_training.setup": {
"total": 0.22507939999999982,
"count": 1,
"self": 0.22507939999999982
},
"TrainerController.start_learning": {
"total": 43718.3264761,
"count": 1,
"self": 6.2553304009270505,
"children": {
"TrainerController._reset_env": {
"total": 7.25175750001204,
"count": 15,
"self": 7.25175750001204
},
"TrainerController.advance": {
"total": 43704.68255109906,
"count": 195046,
"self": 6.6376943998620845,
"children": {
"env_step": {
"total": 27823.66736159923,
"count": 195046,
"self": 26713.80841279826,
"children": {
"SubprocessEnvManager._take_step": {
"total": 1105.5911817002898,
"count": 195046,
"self": 37.86665240134971,
"children": {
"TorchPolicy.evaluate": {
"total": 1067.72452929894,
"count": 386592,
"self": 1067.72452929894
}
}
},
"workers": {
"total": 4.2677671006802465,
"count": 195046,
"self": 0.0,
"children": {
"worker_root": {
"total": 43703.49372489968,
"count": 195046,
"is_parallel": true,
"self": 17770.0365155002,
"children": {
"steps_from_proto": {
"total": 0.05832419999175542,
"count": 30,
"is_parallel": true,
"self": 0.006691800038452378,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.05163239995330304,
"count": 120,
"is_parallel": true,
"self": 0.05163239995330304
}
}
},
"UnityEnvironment.step": {
"total": 25933.398885199487,
"count": 195046,
"is_parallel": true,
"self": 56.72683119988142,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 521.6316486982222,
"count": 195046,
"is_parallel": true,
"self": 521.6316486982222
},
"communicator.exchange": {
"total": 24396.046228599782,
"count": 195046,
"is_parallel": true,
"self": 24396.046228599782
},
"steps_from_proto": {
"total": 958.9941767016019,
"count": 390092,
"is_parallel": true,
"self": 111.15929309924877,
"children": {
"_process_rank_one_or_two_observation": {
"total": 847.8348836023531,
"count": 1560368,
"is_parallel": true,
"self": 847.8348836023531
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 15874.377495099965,
"count": 195046,
"self": 46.30661719941418,
"children": {
"process_trajectory": {
"total": 8990.299759200574,
"count": 195046,
"self": 8988.298889900569,
"children": {
"RLTrainer._checkpoint": {
"total": 2.0008693000054336,
"count": 10,
"self": 2.0008693000054336
}
}
},
"_update_policy": {
"total": 6837.771118699977,
"count": 36,
"self": 692.7114432000171,
"children": {
"TorchPOCAOptimizer.update": {
"total": 6145.05967549996,
"count": 1080,
"self": 6145.05967549996
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.999995770864189e-07,
"count": 1,
"self": 8.999995770864189e-07
},
"TrainerController._save_models": {
"total": 0.13683620000665542,
"count": 1,
"self": 0.0033173000119859353,
"children": {
"RLTrainer._checkpoint": {
"total": 0.13351889999466948,
"count": 1,
"self": 0.13351889999466948
}
}
}
}
}
}
}