{ "Huggy": { "checkpoints": [ { "steps": 199748, "file_path": "results/Huggy/Huggy/Huggy-199748.onnx", "reward": 3.454869327045256, "creation_time": 1672093864.145224, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-199748.pt" ] }, { "steps": 399714, "file_path": "results/Huggy/Huggy/Huggy-399714.onnx", "reward": 3.859131726007613, "creation_time": 1672094081.5939114, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-399714.pt" ] }, { "steps": 599939, "file_path": "results/Huggy/Huggy/Huggy-599939.onnx", "reward": 3.5522091205303488, "creation_time": 1672094303.8853345, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-599939.pt" ] }, { "steps": 799914, "file_path": "results/Huggy/Huggy/Huggy-799914.onnx", "reward": 3.755098972782012, "creation_time": 1672094524.7111545, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-799914.pt" ] }, { "steps": 999867, "file_path": "results/Huggy/Huggy/Huggy-999867.onnx", "reward": 3.6949396394846734, "creation_time": 1672094746.0603642, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-999867.pt" ] }, { "steps": 1199951, "file_path": "results/Huggy/Huggy/Huggy-1199951.onnx", "reward": 4.194747514643912, "creation_time": 1672094968.8830035, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1199951.pt" ] }, { "steps": 1399981, "file_path": "results/Huggy/Huggy/Huggy-1399981.onnx", "reward": null, "creation_time": 1672095193.1705482, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1399981.pt" ] }, { "steps": 1599749, "file_path": "results/Huggy/Huggy/Huggy-1599749.onnx", "reward": 3.680925513783546, "creation_time": 1672095411.431857, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1599749.pt" ] }, { "steps": 1799698, "file_path": "results/Huggy/Huggy/Huggy-1799698.onnx", "reward": 3.4703440873810414, "creation_time": 1672095633.3471231, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1799698.pt" ] }, { "steps": 1999982, "file_path": "results/Huggy/Huggy/Huggy-1999982.onnx", "reward": 3.1711768731474876, "creation_time": 1672095856.690272, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1999982.pt" ] }, { "steps": 2000047, "file_path": "results/Huggy/Huggy/Huggy-2000047.onnx", "reward": 3.1826997434391693, "creation_time": 1672095856.8099647, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000047.pt" ] } ], "final_checkpoint": { "steps": 2000047, "file_path": "results/Huggy/Huggy.onnx", "reward": 3.1826997434391693, "creation_time": 1672095856.8099647, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000047.pt" ] } }, "metadata": { "stats_format_version": "0.3.0", "mlagents_version": "0.29.0.dev0", "torch_version": "1.8.1+cu102" } }