{ "Huggy": { "checkpoints": [ { "steps": 499969, "file_path": "results/Huggy/Huggy/Huggy-499969.onnx", "reward": 4.311801770755223, "creation_time": 1670950266.0459735, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-499969.pt" ] }, { "steps": 999418, "file_path": "results/Huggy/Huggy/Huggy-999418.onnx", "reward": 3.6427455982383417, "creation_time": 1670950846.4277775, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-999418.pt" ] }, { "steps": 1499952, "file_path": "results/Huggy/Huggy/Huggy-1499952.onnx", "reward": 3.5856640924300467, "creation_time": 1670951425.8306668, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1499952.pt" ] }, { "steps": 1999946, "file_path": "results/Huggy/Huggy/Huggy-1999946.onnx", "reward": 4.348629266023636, "creation_time": 1670952006.914395, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1999946.pt" ] }, { "steps": 2000003, "file_path": "results/Huggy/Huggy/Huggy-2000003.onnx", "reward": 4.3680479526519775, "creation_time": 1670952007.0324986, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000003.pt" ] } ], "final_checkpoint": { "steps": 2000003, "file_path": "results/Huggy/Huggy.onnx", "reward": 4.3680479526519775, "creation_time": 1670952007.0324986, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000003.pt" ] } }, "metadata": { "stats_format_version": "0.3.0", "mlagents_version": "0.29.0.dev0", "torch_version": "1.8.1+cu102" } }