sgoodfriend's picture
VPG playing BipedalWalker-v3 from https://github.com/sgoodfriend/rl-algo-impls/tree/0511de345b17175b7cf1ea706c3e05981f11761c
a902404
import os
from typing import Optional, Sequence, TypeVar
import numpy as np
import torch
from rl_algo_impls.dqn.q_net import QNetwork
from rl_algo_impls.shared.policy.policy import Policy
from rl_algo_impls.wrappers.vectorable_wrapper import (
VecEnv,
VecEnvObs,
single_action_space,
single_observation_space,
)
DQNPolicySelf = TypeVar("DQNPolicySelf", bound="DQNPolicy")
class DQNPolicy(Policy):
def __init__(
self,
env: VecEnv,
hidden_sizes: Sequence[int] = [],
cnn_flatten_dim: int = 512,
cnn_style: str = "nature",
cnn_layers_init_orthogonal: Optional[bool] = None,
impala_channels: Sequence[int] = (16, 32, 32),
**kwargs,
) -> None:
super().__init__(env, **kwargs)
self.q_net = QNetwork(
single_observation_space(env),
single_action_space(env),
hidden_sizes,
cnn_flatten_dim=cnn_flatten_dim,
cnn_style=cnn_style,
cnn_layers_init_orthogonal=cnn_layers_init_orthogonal,
impala_channels=impala_channels,
)
def act(
self,
obs: VecEnvObs,
eps: float = 0,
deterministic: bool = True,
action_masks: Optional[np.ndarray] = None,
) -> np.ndarray:
assert eps == 0 if deterministic else eps >= 0
assert (
action_masks is None
), f"action_masks not currently supported in {self.__class__.__name__}"
if not deterministic and np.random.random() < eps:
return np.array(
[
single_action_space(self.env).sample()
for _ in range(self.env.num_envs)
]
)
else:
o = self._as_tensor(obs)
with torch.no_grad():
return self.q_net(o).argmax(axis=1).cpu().numpy()