colinrgodsey commited on
Commit
07f570f
1 Parent(s): d60795c

First Push

Browse files
Pyramids-193860.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26ff8f4e4830c359b54027532484932435aff33f008334be9182d9588da27a3d
3
+ size 8656173
README.md ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - Pyramids
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-Pyramids
8
+ ---
9
+
10
+ # **ppo** Agent playing **Pyramids**
11
+ This is a trained model of a **ppo** agent playing **Pyramids**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ ## Usage (with ML-Agents)
15
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
+
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+ - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
+ - A *longer tutorial* to understand how works ML-Agents:
21
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
+
23
+ ### Resume the training
24
+ ```bash
25
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
+ ```
27
+
28
+ ### Watch your Agent play
29
+ You can watch your agent **playing directly in your browser**
30
+
31
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
+ 2. Step 1: Find your model_id: colinrgodsey/ml-agents-pyramid
33
+ 3. Step 2: Select your *.nn /*.onnx file
34
+ 4. Click on Watch the agent play 👀
35
+
checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9121a78ea098e4be139afb389a2064f40bbd1327dd5b237e7d7eabf0d85733cd
3
+ size 8655778
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple"}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"hidden_units": 64, "num_layers": 3}, "learning_rate": 0.0001}}, "keep_checkpoints": 5, "max_steps": 3000000, "time_horizon": 128, "summary_freq": 30000}}}
configuration.yaml ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ behaviors:
2
+ Pyramids:
3
+ trainer_type: ppo
4
+ hyperparameters:
5
+ batch_size: 128
6
+ buffer_size: 2048
7
+ learning_rate: 0.0003
8
+ beta: 0.01
9
+ epsilon: 0.2
10
+ lambd: 0.95
11
+ num_epoch: 3
12
+ learning_rate_schedule: linear
13
+ network_settings:
14
+ normalize: false
15
+ hidden_units: 512
16
+ num_layers: 2
17
+ vis_encode_type: simple
18
+ reward_signals:
19
+ extrinsic:
20
+ gamma: 0.99
21
+ strength: 1.0
22
+ rnd:
23
+ gamma: 0.99
24
+ strength: 0.01
25
+ network_settings:
26
+ hidden_units: 64
27
+ num_layers: 3
28
+ learning_rate: 0.0001
29
+ keep_checkpoints: 5
30
+ max_steps: 3000000
31
+ time_horizon: 128
32
+ summary_freq: 30000
events.out.tfevents.1722712150.tf-notebook-0.61766.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26074d5db9339ed5347d247f7145940aebc81e450616ce744eb403b5df830dfc
3
+ size 16380