Shortcuts

GAE

class torchrl.objectives.value.GAE(*args, **kwargs)[source]

A class wrapper around the generalized advantage estimate functional.

Refer to “HIGH-DIMENSIONAL CONTINUOUS CONTROL USING GENERALIZED ADVANTAGE ESTIMATION” https://arxiv.org/pdf/1506.02438.pdf for more context.

Parameters:
  • gamma (scalar) – exponential mean discount.

  • lmbda (scalar) – trajectory discount.

  • value_network (TensorDictModule) – value operator used to retrieve the value estimates.

  • average_gae (bool) – if True, the resulting GAE values will be standardized. Default is False.

  • differentiable (bool, optional) –

    if True, gradients are propagated through the computation of the value function. Default is False.

    Note

    The proper way to make the function call non-differentiable is to decorate it in a torch.no_grad() context manager/decorator or pass detached parameters for functional modules.

  • vectorized (bool, optional) – whether to use the vectorized version of the lambda return. Default is True if not compiling.

  • skip_existing (bool, optional) – if True, the value network will skip modules which outputs are already present in the tensordict. Defaults to None, i.e., the value of tensordict.nn.skip_existing() is not affected. Defaults to “state_value”.

  • advantage_key (str or tuple of str, optional) – [Deprecated] the key of the advantage entry. Defaults to "advantage".

  • value_target_key (str or tuple of str, optional) – [Deprecated] the key of the advantage entry. Defaults to "value_target".

  • value_key (str or tuple of str, optional) – [Deprecated] the value key to read from the input tensordict. Defaults to "state_value".

  • shifted (bool, optional) – if True, the value and next value are estimated with a single call to the value network. This is faster but is only valid whenever (1) the "next" value is shifted by only one time step (which is not the case with multi-step value estimation, for instance) and (2) when the parameters used at time t and t+1 are identical (which is not the case when target parameters are to be used). Defaults to False.

  • device (torch.device, optional) – the device where the buffers will be instantiated. Defaults to torch.get_default_device().

  • time_dim (int, optional) – the dimension corresponding to the time in the input tensordict. If not provided, defaults to the dimension marked with the "time" name if any, and to the last dimension otherwise. Can be overridden during a call to value_estimate(). Negative dimensions are considered with respect to the input tensordict.

  • auto_reset_env (bool, optional) – if True, the last "next" state of the episode isn’t valid, so the GAE calculation will use the value instead of next_value to bootstrap truncated episodes.

  • deactivate_vmap (bool, optional) – if True, no vmap call will be used, and vectorized maps will be replaced with simple for loops. Defaults to False.

GAE will return an "advantage" entry containing the advantage value. It will also return a "value_target" entry with the return value that is to be used to train the value network. Finally, if gradient_mode is True, an additional and differentiable "value_error" entry will be returned, which simply represents the difference between the return and the value network output (i.e. an additional distance loss should be applied to that signed value).

Note

As other advantage functions do, if the value_key is already present in the input tensordict, the GAE module will ignore the calls to the value network (if any) and use the provided value instead.

Note

GAE can be used with value networks that rely on recurrent neural networks, provided that the init markers (“is_init”) and terminated / truncated markers are properly set. If shifted=True, the trajectory batch will be flattened and the last step of each trajectory will be placed within the flat tensordict after the last step from the root, such that each trajectory has T+1 elements. If shifted=False, the root and “next” trajecotries will be stacked and the value network will be called with vmap over the stack of trajectories. Because RNNs require fair amount of control flow, they are currently not compatible with torch.vmap and, as such, the deactivate_vmap option must be turned on in these cases. Similarly, if shifted=False, the “is_init” entry of the root tensordict will be copied onto the “is_init” of the “next” entry, such that trajectories are well separated both for root and “next” data.

forward(tensordict=None, *, params: list[Tensor] | None = None, target_params: list[Tensor] | None = None, time_dim: int | None = None)[source]

Computes the GAE given the data in tensordict.

If a functional module is provided, a nested TensorDict containing the parameters (and if relevant the target parameters) can be passed to the module.

Parameters:

tensordict (TensorDictBase) – A TensorDict containing the data (an observation key, "action", ("next", "reward"), ("next", "done"), ("next", "terminated"), and "next" tensordict state as returned by the environment) necessary to compute the value estimates and the GAE. The data passed to this module should be structured as [*B, T, *F] where B are the batch size, T the time dimension and F the feature dimension(s). The tensordict must have shape [*B, T].

Keyword Arguments:
  • params (TensorDictBase, optional) – A nested TensorDict containing the params to be passed to the functional value network module.

  • target_params (TensorDictBase, optional) – A nested TensorDict containing the target params to be passed to the functional value network module.

  • time_dim (int, optional) – the dimension corresponding to the time in the input tensordict. If not provided, defaults to the dimension marked with the "time" name if any, and to the last dimension otherwise. Negative dimensions are considered with respect to the input tensordict.

Returns:

An updated TensorDict with an advantage and a value_error keys as defined in the constructor.

Examples

>>> from tensordict import TensorDict
>>> value_net = TensorDictModule(
...     nn.Linear(3, 1), in_keys=["obs"], out_keys=["state_value"]
... )
>>> module = GAE(
...     gamma=0.98,
...     lmbda=0.94,
...     value_network=value_net,
...     differentiable=False,
... )
>>> obs, next_obs = torch.randn(2, 1, 10, 3)
>>> reward = torch.randn(1, 10, 1)
>>> done = torch.zeros(1, 10, 1, dtype=torch.bool)
>>> terminated = torch.zeros(1, 10, 1, dtype=torch.bool)
>>> tensordict = TensorDict({"obs": obs, "next": {"obs": next_obs}, "done": done, "reward": reward, "terminated": terminated}, [1, 10])
>>> _ = module(tensordict)
>>> assert "advantage" in tensordict.keys()

The module supports non-tensordict (i.e. unpacked tensordict) inputs too:

Examples

>>> value_net = TensorDictModule(
...     nn.Linear(3, 1), in_keys=["obs"], out_keys=["state_value"]
... )
>>> module = GAE(
...     gamma=0.98,
...     lmbda=0.94,
...     value_network=value_net,
...     differentiable=False,
... )
>>> obs, next_obs = torch.randn(2, 1, 10, 3)
>>> reward = torch.randn(1, 10, 1)
>>> done = torch.zeros(1, 10, 1, dtype=torch.bool)
>>> terminated = torch.zeros(1, 10, 1, dtype=torch.bool)
>>> advantage, value_target = module(obs=obs, next_reward=reward, next_done=done, next_obs=next_obs, next_terminated=terminated)
value_estimate(tensordict, params: TensorDictBase | None = None, target_params: TensorDictBase | None = None, time_dim: int | None = None, **kwargs)[source]

Gets a value estimate, usually used as a target value for the value network.

If the state value key is present under tensordict.get(("next", self.tensor_keys.value)) then this value will be used without recurring to the value network.

Parameters:
  • tensordict (TensorDictBase) – the tensordict containing the data to read.

  • target_params (TensorDictBase, optional) – A nested TensorDict containing the target params to be passed to the functional value network module.

  • next_value (torch.Tensor, optional) – the value of the next state or state-action pair. Exclusive with target_params.

  • **kwargs – the keyword arguments to be passed to the value network.

Returns: a tensor corresponding to the state value.

Docs

Access comprehensive developer documentation for PyTorch

View Docs

Tutorials

Get in-depth tutorials for beginners and advanced developers

View Tutorials

Resources

Find development resources and get your questions answered

View Resources