{ "Huggy": { "checkpoints": [ { "steps": 499958, "file_path": "results/Huggy/Huggy/Huggy-499958.onnx", "reward": 4.289363158054841, "creation_time": 1670593031.0694492, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-499958.pt" ] }, { "steps": 999999, "file_path": "results/Huggy/Huggy/Huggy-999999.onnx", "reward": 3.6988724134862423, "creation_time": 1670593569.6651065, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-999999.pt" ] }, { "steps": 1499523, "file_path": "results/Huggy/Huggy/Huggy-1499523.onnx", "reward": 3.7572394197035317, "creation_time": 1670594111.8129506, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1499523.pt" ] }, { "steps": 1999945, "file_path": "results/Huggy/Huggy/Huggy-1999945.onnx", "reward": 3.144189735253652, "creation_time": 1670594655.7657974, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1999945.pt" ] }, { "steps": 2000013, "file_path": "results/Huggy/Huggy/Huggy-2000013.onnx", "reward": 3.1715092847221777, "creation_time": 1670594655.8836553, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000013.pt" ] } ], "final_checkpoint": { "steps": 2000013, "file_path": "results/Huggy/Huggy.onnx", "reward": 3.1715092847221777, "creation_time": 1670594655.8836553, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000013.pt" ] } }, "metadata": { "stats_format_version": "0.3.0", "mlagents_version": "0.29.0.dev0", "torch_version": "1.8.1+cu102" } }