mirror of
https://github.com/saymrwulf/stable-baselines3.git
synced 2026-05-15 21:00:53 +00:00
* Add support for custom objects * Add python 3.8 to the CI * Bump version * PyType fixes * [ci skip] Fix typo * Add note about slow-down + fix typos * Minor edits to the doc * Bug fix for DQN * Update test * Add test for custom objects
165 lines
4.3 KiB
ReStructuredText
165 lines
4.3 KiB
ReStructuredText
.. _her:
|
|
|
|
.. automodule:: stable_baselines3.her
|
|
|
|
|
|
HER
|
|
====
|
|
|
|
`Hindsight Experience Replay (HER) <https://arxiv.org/abs/1707.01495>`_
|
|
|
|
HER is an algorithm that works with off-policy methods (DQN, SAC, TD3 and DDPG for example).
|
|
HER uses the fact that even if a desired goal was not achieved, other goal may have been achieved during a rollout.
|
|
It creates "virtual" transitions by relabeling transitions (changing the desired goal) from past episodes.
|
|
|
|
|
|
|
|
.. warning::
|
|
|
|
HER requires the environment to inherits from `gym.GoalEnv <https://github.com/openai/gym/blob/3394e245727c1ae6851b504a50ba77c73cd4c65b/gym/core.py#L160>`_
|
|
|
|
|
|
.. warning::
|
|
|
|
For performance reasons, the maximum number of steps per episodes must be specified.
|
|
In most cases, it will be inferred if you specify ``max_episode_steps`` when registering the environment
|
|
or if you use a ``gym.wrappers.TimeLimit`` (and ``env.spec`` is not None).
|
|
Otherwise, you can directly pass ``max_episode_length`` to the model constructor
|
|
|
|
|
|
.. warning::
|
|
|
|
``HER`` supports ``VecNormalize`` wrapper but only when ``online_sampling=True``
|
|
|
|
|
|
.. warning::
|
|
|
|
Because it needs access to ``env.compute_reward()``
|
|
``HER`` must be loaded with the env. If you just want to use the trained policy
|
|
without instantiating the environment, we recommend saving the policy only.
|
|
|
|
|
|
Notes
|
|
-----
|
|
|
|
- Original paper: https://arxiv.org/abs/1707.01495
|
|
- OpenAI paper: `Plappert et al. (2018)`_
|
|
- OpenAI blog post: https://openai.com/blog/ingredients-for-robotics-research/
|
|
|
|
|
|
.. _Plappert et al. (2018): https://arxiv.org/abs/1802.09464
|
|
|
|
Can I use?
|
|
----------
|
|
|
|
Please refer to the used model (DQN, QR-DQN, SAC, TQC, TD3, or DDPG) for that section.
|
|
|
|
Example
|
|
-------
|
|
|
|
.. code-block:: python
|
|
|
|
from stable_baselines3 import HER, DDPG, DQN, SAC, TD3
|
|
from stable_baselines3.her.goal_selection_strategy import GoalSelectionStrategy
|
|
from stable_baselines3.common.bit_flipping_env import BitFlippingEnv
|
|
from stable_baselines3.common.vec_env import DummyVecEnv
|
|
from stable_baselines3.common.vec_env.obs_dict_wrapper import ObsDictWrapper
|
|
|
|
model_class = DQN # works also with SAC, DDPG and TD3
|
|
N_BITS = 15
|
|
|
|
env = BitFlippingEnv(n_bits=N_BITS, continuous=model_class in [DDPG, SAC, TD3], max_steps=N_BITS)
|
|
|
|
# Available strategies (cf paper): future, final, episode
|
|
goal_selection_strategy = 'future' # equivalent to GoalSelectionStrategy.FUTURE
|
|
|
|
# If True the HER transitions will get sampled online
|
|
online_sampling = True
|
|
# Time limit for the episodes
|
|
max_episode_length = N_BITS
|
|
|
|
# Initialize the model
|
|
model = HER('MlpPolicy', env, model_class, n_sampled_goal=4, goal_selection_strategy=goal_selection_strategy, online_sampling=online_sampling,
|
|
verbose=1, max_episode_length=max_episode_length)
|
|
# Train the model
|
|
model.learn(1000)
|
|
|
|
model.save("./her_bit_env")
|
|
# Because it needs access to `env.compute_reward()`
|
|
# HER must be loaded with the env
|
|
model = HER.load('./her_bit_env', env=env)
|
|
|
|
obs = env.reset()
|
|
for _ in range(100):
|
|
action, _ = model.predict(obs, deterministic=True)
|
|
obs, reward, done, _ = env.step(action)
|
|
|
|
if done:
|
|
obs = env.reset()
|
|
|
|
|
|
Results
|
|
-------
|
|
|
|
This implementation was tested on the `parking env <https://github.com/eleurent/highway-env>`_
|
|
using 3 seeds.
|
|
|
|
The complete learning curves are available in the `associated PR #120 <https://github.com/DLR-RM/stable-baselines3/pull/120>`_.
|
|
|
|
|
|
|
|
How to replicate the results?
|
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
Clone the `rl-zoo repo <https://github.com/DLR-RM/rl-baselines3-zoo>`_:
|
|
|
|
.. code-block:: bash
|
|
|
|
git clone https://github.com/DLR-RM/rl-baselines3-zoo
|
|
cd rl-baselines3-zoo/
|
|
|
|
|
|
Run the benchmark:
|
|
|
|
.. code-block:: bash
|
|
|
|
python train.py --algo her --env parking-v0 --eval-episodes 10 --eval-freq 10000
|
|
|
|
|
|
Plot the results:
|
|
|
|
.. code-block:: bash
|
|
|
|
python scripts/all_plots.py -a her -e parking-v0 -f logs/ --no-million
|
|
|
|
|
|
Parameters
|
|
----------
|
|
|
|
.. autoclass:: HER
|
|
:members:
|
|
|
|
Goal Selection Strategies
|
|
-------------------------
|
|
|
|
.. autoclass:: GoalSelectionStrategy
|
|
:members:
|
|
:inherited-members:
|
|
:undoc-members:
|
|
|
|
|
|
Obs Dict Wrapper
|
|
----------------
|
|
|
|
.. autoclass:: ObsDictWrapper
|
|
:members:
|
|
:inherited-members:
|
|
:undoc-members:
|
|
|
|
|
|
HER Replay Buffer
|
|
-----------------
|
|
|
|
.. autoclass:: HerReplayBuffer
|
|
:members:
|
|
:inherited-members:
|