|
!!python/object/apply:collections.OrderedDict |
|
- - - batch_size |
|
- 256 |
|
- - clip_range |
|
- 0.2 |
|
- - ent_coef |
|
- 2.0745206045994986e-05 |
|
- - gae_lambda |
|
- 0.92 |
|
- - gamma |
|
- 0.999 |
|
- - learning_rate |
|
- 2.0309225666232827e-05 |
|
- - max_grad_norm |
|
- 0.5 |
|
- - n_envs |
|
- 1 |
|
- - n_epochs |
|
- 20 |
|
- - n_steps |
|
- 2048 |
|
- - n_timesteps |
|
- 10000000.0 |
|
- - normalize |
|
- gamma: 0.999 |
|
norm_obs: false |
|
norm_reward: true |
|
- - policy |
|
- MlpPolicy |
|
- - policy_kwargs |
|
- activation_fn: !!python/name:torch.nn.modules.activation.ReLU '' |
|
features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor '' |
|
net_arch: |
|
- pi: |
|
- 256 |
|
- 256 |
|
vf: |
|
- 256 |
|
- 256 |
|
- - vf_coef |
|
- 0.819262464558427 |
|
|