2020-05-12 11:31:47 +08:00
|
|
|
import torch
|
2020-04-14 21:11:06 +08:00
|
|
|
import numpy as np
|
2020-03-18 21:45:41 +08:00
|
|
|
from torch import nn
|
2020-03-12 22:20:33 +08:00
|
|
|
from abc import ABC, abstractmethod
|
2020-06-02 22:29:50 +08:00
|
|
|
from typing import Dict, List, Union, Optional, Callable
|
2020-05-12 11:31:47 +08:00
|
|
|
|
2020-06-03 13:59:47 +08:00
|
|
|
from tianshou.data import Batch, ReplayBuffer, to_torch_as
|
2020-03-12 22:20:33 +08:00
|
|
|
|
|
|
|
|
2020-03-18 21:45:41 +08:00
|
|
|
class BasePolicy(ABC, nn.Module):
|
2020-04-06 19:36:59 +08:00
|
|
|
"""Tianshou aims to modularizing RL algorithms. It comes into several
|
|
|
|
classes of policies in Tianshou. All of the policy classes must inherit
|
|
|
|
:class:`~tianshou.policy.BasePolicy`.
|
2020-03-13 17:49:22 +08:00
|
|
|
|
2020-04-06 19:36:59 +08:00
|
|
|
A policy class typically has four parts:
|
|
|
|
|
|
|
|
* :meth:`~tianshou.policy.BasePolicy.__init__`: initialize the policy, \
|
|
|
|
including coping the target network and so on;
|
2020-04-10 10:47:16 +08:00
|
|
|
* :meth:`~tianshou.policy.BasePolicy.forward`: compute action with given \
|
2020-04-06 19:36:59 +08:00
|
|
|
observation;
|
|
|
|
* :meth:`~tianshou.policy.BasePolicy.process_fn`: pre-process data from \
|
|
|
|
the replay buffer (this function can interact with replay buffer);
|
|
|
|
* :meth:`~tianshou.policy.BasePolicy.learn`: update policy with a given \
|
|
|
|
batch of data.
|
|
|
|
|
|
|
|
Most of the policy needs a neural network to predict the action and an
|
|
|
|
optimizer to optimize the policy. The rules of self-defined networks are:
|
|
|
|
|
2020-06-08 21:53:00 +08:00
|
|
|
1. Input: observation ``obs`` (may be a ``numpy.ndarray``, a \
|
|
|
|
``torch.Tensor``, a dict or any others), hidden state ``state`` (for \
|
|
|
|
RNN usage), and other information ``info`` provided by the \
|
|
|
|
environment.
|
|
|
|
2. Output: some ``logits``, the next hidden state ``state``, and the \
|
|
|
|
intermediate result during policy forwarding procedure ``policy``. The\
|
|
|
|
``logits`` could be a tuple instead of a ``torch.Tensor``. It depends \
|
|
|
|
on how the policy process the network output. For example, in PPO, the\
|
|
|
|
return of the network might be ``(mu, sigma), state`` for Gaussian \
|
|
|
|
policy. The ``policy`` can be a Batch of torch.Tensor or other things,\
|
|
|
|
which will be stored in the replay buffer, and can be accessed in the \
|
|
|
|
policy update process (e.g. in ``policy.learn()``, the \
|
|
|
|
``batch.policy`` is what you need).
|
2020-04-06 19:36:59 +08:00
|
|
|
|
|
|
|
Since :class:`~tianshou.policy.BasePolicy` inherits ``torch.nn.Module``,
|
2020-04-10 11:16:33 +08:00
|
|
|
you can use :class:`~tianshou.policy.BasePolicy` almost the same as
|
|
|
|
``torch.nn.Module``, for instance, loading and saving the model:
|
2020-04-06 19:36:59 +08:00
|
|
|
::
|
|
|
|
|
|
|
|
torch.save(policy.state_dict(), 'policy.pth')
|
|
|
|
policy.load_state_dict(torch.load('policy.pth'))
|
|
|
|
"""
|
|
|
|
|
2020-05-12 11:31:47 +08:00
|
|
|
def __init__(self, **kwargs) -> None:
|
2020-03-12 22:20:33 +08:00
|
|
|
super().__init__()
|
2020-06-09 18:46:14 +08:00
|
|
|
self.observation_space = kwargs.get('observation_space')
|
|
|
|
self.action_space = kwargs.get('action_space')
|
2020-03-12 22:20:33 +08:00
|
|
|
|
2020-05-12 11:31:47 +08:00
|
|
|
def process_fn(self, batch: Batch, buffer: ReplayBuffer,
|
|
|
|
indice: np.ndarray) -> Batch:
|
2020-04-06 19:36:59 +08:00
|
|
|
"""Pre-process the data from the provided replay buffer. Check out
|
|
|
|
:ref:`policy_concept` for more information.
|
|
|
|
"""
|
2020-03-17 11:37:31 +08:00
|
|
|
return batch
|
|
|
|
|
2020-03-12 22:20:33 +08:00
|
|
|
@abstractmethod
|
2020-05-12 11:31:47 +08:00
|
|
|
def forward(self, batch: Batch,
|
|
|
|
state: Optional[Union[dict, Batch, np.ndarray]] = None,
|
|
|
|
**kwargs) -> Batch:
|
2020-04-06 19:36:59 +08:00
|
|
|
"""Compute action over the given batch data.
|
|
|
|
|
|
|
|
:return: A :class:`~tianshou.data.Batch` which MUST have the following\
|
|
|
|
keys:
|
|
|
|
|
|
|
|
* ``act`` an numpy.ndarray or a torch.Tensor, the action over \
|
|
|
|
given batch data.
|
|
|
|
* ``state`` a dict, an numpy.ndarray or a torch.Tensor, the \
|
|
|
|
internal state of the policy, ``None`` as default.
|
|
|
|
|
|
|
|
Other keys are user-defined. It depends on the algorithm. For example,
|
|
|
|
::
|
|
|
|
|
|
|
|
# some code
|
|
|
|
return Batch(logits=..., act=..., state=None, dist=...)
|
2020-04-29 17:48:48 +08:00
|
|
|
|
|
|
|
After version >= 0.2.3, the keyword "policy" is reserverd and the
|
|
|
|
corresponding data will be stored into the replay buffer in numpy. For
|
|
|
|
instance,
|
|
|
|
::
|
|
|
|
|
|
|
|
# some code
|
|
|
|
return Batch(..., policy=Batch(log_prob=dist.log_prob(act)))
|
|
|
|
# and in the sampled data batch, you can directly call
|
|
|
|
# batch.policy.log_prob to get your data, although it is stored in
|
|
|
|
# np.ndarray.
|
2020-04-06 19:36:59 +08:00
|
|
|
"""
|
2020-03-12 22:20:33 +08:00
|
|
|
pass
|
|
|
|
|
2020-03-15 17:41:00 +08:00
|
|
|
@abstractmethod
|
2020-05-12 11:31:47 +08:00
|
|
|
def learn(self, batch: Batch, **kwargs
|
|
|
|
) -> Dict[str, Union[float, List[float]]]:
|
2020-04-06 19:36:59 +08:00
|
|
|
"""Update policy with a given batch of data.
|
2020-03-12 22:20:33 +08:00
|
|
|
|
2020-04-06 19:36:59 +08:00
|
|
|
:return: A dict which includes loss and its corresponding label.
|
|
|
|
"""
|
2020-03-14 21:48:31 +08:00
|
|
|
pass
|
2020-04-14 21:11:06 +08:00
|
|
|
|
2020-04-19 14:30:42 +08:00
|
|
|
@staticmethod
|
2020-05-12 11:31:47 +08:00
|
|
|
def compute_episodic_return(
|
|
|
|
batch: Batch,
|
|
|
|
v_s_: Optional[Union[np.ndarray, torch.Tensor]] = None,
|
2020-05-16 20:08:32 +08:00
|
|
|
gamma: float = 0.99,
|
2020-07-16 19:36:32 +08:00
|
|
|
gae_lambda: float = 0.95,
|
|
|
|
) -> Batch:
|
2020-04-14 21:11:06 +08:00
|
|
|
"""Compute returns over given full-length episodes, including the
|
2020-05-27 11:02:23 +08:00
|
|
|
implementation of Generalized Advantage Estimator (arXiv:1506.02438).
|
2020-04-14 21:11:06 +08:00
|
|
|
|
|
|
|
:param batch: a data batch which contains several full-episode data
|
|
|
|
chronologically.
|
|
|
|
:type batch: :class:`~tianshou.data.Batch`
|
|
|
|
:param v_s_: the value function of all next states :math:`V(s')`.
|
|
|
|
:type v_s_: numpy.ndarray
|
|
|
|
:param float gamma: the discount factor, should be in [0, 1], defaults
|
|
|
|
to 0.99.
|
|
|
|
:param float gae_lambda: the parameter for Generalized Advantage
|
|
|
|
Estimation, should be in [0, 1], defaults to 0.95.
|
2020-06-02 22:29:50 +08:00
|
|
|
|
|
|
|
:return: a Batch. The result will be stored in batch.returns.
|
2020-04-14 21:11:06 +08:00
|
|
|
"""
|
2020-07-16 19:36:32 +08:00
|
|
|
rew = batch.rew
|
2020-04-14 21:11:06 +08:00
|
|
|
if v_s_ is None:
|
2020-07-16 19:36:32 +08:00
|
|
|
v_s_ = rew * 0.
|
2020-04-14 21:11:06 +08:00
|
|
|
else:
|
2020-04-19 14:30:42 +08:00
|
|
|
if not isinstance(v_s_, np.ndarray):
|
|
|
|
v_s_ = np.array(v_s_, np.float)
|
2020-07-16 19:36:32 +08:00
|
|
|
v_s_ = v_s_.reshape(rew.shape)
|
2020-06-02 22:29:50 +08:00
|
|
|
returns = np.roll(v_s_, 1, axis=0)
|
2020-04-14 21:11:06 +08:00
|
|
|
m = (1. - batch.done) * gamma
|
2020-07-16 19:36:32 +08:00
|
|
|
delta = rew + v_s_ * m - returns
|
2020-04-14 21:11:06 +08:00
|
|
|
m *= gae_lambda
|
|
|
|
gae = 0.
|
2020-07-16 19:36:32 +08:00
|
|
|
for i in range(len(rew) - 1, -1, -1):
|
2020-04-14 21:11:06 +08:00
|
|
|
gae = delta[i] + m[i] * gae
|
2020-06-02 22:29:50 +08:00
|
|
|
returns[i] += gae
|
|
|
|
batch.returns = returns
|
|
|
|
return batch
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def compute_nstep_return(
|
|
|
|
batch: Batch,
|
|
|
|
buffer: ReplayBuffer,
|
|
|
|
indice: np.ndarray,
|
2020-06-03 13:59:47 +08:00
|
|
|
target_q_fn: Callable[[ReplayBuffer, np.ndarray], torch.Tensor],
|
2020-06-02 22:29:50 +08:00
|
|
|
gamma: float = 0.99,
|
2020-06-03 13:59:47 +08:00
|
|
|
n_step: int = 1,
|
2020-07-16 19:36:32 +08:00
|
|
|
rew_norm: bool = False,
|
2020-06-02 22:29:50 +08:00
|
|
|
) -> np.ndarray:
|
|
|
|
r"""Compute n-step return for Q-learning targets:
|
|
|
|
|
|
|
|
.. math::
|
|
|
|
G_t = \sum_{i = t}^{t + n - 1} \gamma^{i - t}(1 - d_i)r_i +
|
|
|
|
\gamma^n (1 - d_{t + n}) Q_{\mathrm{target}}(s_{t + n})
|
|
|
|
|
|
|
|
, where :math:`\gamma` is the discount factor,
|
|
|
|
:math:`\gamma \in [0, 1]`, :math:`d_t` is the done flag of step
|
|
|
|
:math:`t`.
|
|
|
|
|
|
|
|
:param batch: a data batch, which is equal to buffer[indice].
|
|
|
|
:type batch: :class:`~tianshou.data.Batch`
|
|
|
|
:param buffer: a data buffer which contains several full-episode data
|
|
|
|
chronologically.
|
|
|
|
:type buffer: :class:`~tianshou.data.ReplayBuffer`
|
|
|
|
:param indice: sampled timestep.
|
|
|
|
:type indice: numpy.ndarray
|
2020-06-03 13:59:47 +08:00
|
|
|
:param function target_q_fn: a function receives :math:`t+n-1` step's
|
|
|
|
data and compute target Q value.
|
2020-06-02 22:29:50 +08:00
|
|
|
:param float gamma: the discount factor, should be in [0, 1], defaults
|
|
|
|
to 0.99.
|
|
|
|
:param int n_step: the number of estimation step, should be an int
|
|
|
|
greater than 0, defaults to 1.
|
2020-06-03 13:59:47 +08:00
|
|
|
:param bool rew_norm: normalize the reward to Normal(0, 1), defaults
|
|
|
|
to ``False``.
|
2020-06-02 22:29:50 +08:00
|
|
|
|
2020-06-03 13:59:47 +08:00
|
|
|
:return: a Batch. The result will be stored in batch.returns as a
|
|
|
|
torch.Tensor with shape (bsz, ).
|
2020-06-02 22:29:50 +08:00
|
|
|
"""
|
2020-07-16 19:36:32 +08:00
|
|
|
rew = buffer.rew
|
2020-06-03 13:59:47 +08:00
|
|
|
if rew_norm:
|
2020-07-16 19:36:32 +08:00
|
|
|
bfr = rew[:min(len(buffer), 1000)] # avoid large buffer
|
2020-06-03 13:59:47 +08:00
|
|
|
mean, std = bfr.mean(), bfr.std()
|
|
|
|
if np.isclose(std, 0):
|
|
|
|
mean, std = 0, 1
|
|
|
|
else:
|
|
|
|
mean, std = 0, 1
|
2020-06-02 22:29:50 +08:00
|
|
|
returns = np.zeros_like(indice)
|
|
|
|
gammas = np.zeros_like(indice) + n_step
|
2020-07-16 19:36:32 +08:00
|
|
|
done, buf_len = buffer.done, len(buffer)
|
2020-06-02 22:29:50 +08:00
|
|
|
for n in range(n_step - 1, -1, -1):
|
|
|
|
now = (indice + n) % buf_len
|
|
|
|
gammas[done[now] > 0] = n
|
|
|
|
returns[done[now] > 0] = 0
|
2020-06-03 13:59:47 +08:00
|
|
|
returns = (rew[now] - mean) / std + gamma * returns
|
2020-06-02 22:29:50 +08:00
|
|
|
terminal = (indice + n_step - 1) % buf_len
|
2020-06-03 13:59:47 +08:00
|
|
|
target_q = target_q_fn(buffer, terminal).squeeze()
|
2020-06-02 22:29:50 +08:00
|
|
|
target_q[gammas != n_step] = 0
|
2020-06-03 13:59:47 +08:00
|
|
|
returns = to_torch_as(returns, target_q)
|
|
|
|
gammas = to_torch_as(gamma ** gammas, target_q)
|
|
|
|
batch.returns = target_q * gammas + returns
|
2020-04-14 21:11:06 +08:00
|
|
|
return batch
|