ppo-seals-Humanoid-v1 / config.yml
ernestum's picture
Initial commit
f0612ed
raw
history blame contribute delete
815 Bytes
!!python/object/apply:collections.OrderedDict
- - - batch_size
- 256
- - clip_range
- 0.2
- - ent_coef
- 2.0745206045994986e-05
- - gae_lambda
- 0.92
- - gamma
- 0.999
- - learning_rate
- 2.0309225666232827e-05
- - max_grad_norm
- 0.5
- - n_envs
- 1
- - n_epochs
- 20
- - n_steps
- 2048
- - n_timesteps
- 10000000.0
- - normalize
- gamma: 0.999
norm_obs: false
norm_reward: true
- - policy
- MlpPolicy
- - policy_kwargs
- activation_fn: !!python/name:torch.nn.modules.activation.ReLU ''
features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor ''
net_arch:
- pi:
- 256
- 256
vf:
- 256
- 256
- - vf_coef
- 0.819262464558427