mirror of
https://github.com/saymrwulf/stable-baselines3.git
synced 2026-05-17 21:20:11 +00:00
* Created DQN template according to the paper. Next steps: - Create Policy - Complete Training - Debug * Changed Base Class * refactor save, to be consistence with overriding the excluded_save_params function. Do not try to exclude the parameters twice. * Added simple DQN policy * Finished learn and train function - missing correct loss computation * changed collect_rollouts to work with discrete space * moved discrete space collect_rollouts to dqn * basic dqn working * deleted SDE related code * added gradient clipping and moved greedy policy to policy * changed policy to implement target network and added soft update(in fact standart tau is 1 so hard update) * fixed policy setup * rebase target_update_intervall on _n_updates * adapted all tests all tests passing * Move to stable-baseline3 * Fixes for DQN * Fix tests + add CNNPolicy * Allow any optimizer for DQN * added some util functions to create a arbitrary linear schedule, fixed pickle problem with old exploration schedule * more documentation * changed buffer dtype * refactor and document * Added Sphinx Documentation Updated changelog.rst * removed custom collect_rollouts as it is no longer necessary * Implemented suggestions to clean code and documentation. * extracted some functions on tests to reduce duplicated code * added support for exploration_fraction * Fixed exploration_fraction * Added documentation * Fixed get_linear_fn -> proper progress scaling * Merged master * Added nature reference * Changed default parameters to https://www.nature.com/articles/nature14236/tables/1 * Fixed n_updates to be incremented correctly * Correct train_freq * Doc update * added special parameter for DQN in tests * different fix for test_discrete * Update docs/modules/dqn.rst Co-authored-by: Antonin RAFFIN <antonin.raffin@ensta.org> * Update docs/modules/dqn.rst Co-authored-by: Antonin RAFFIN <antonin.raffin@ensta.org> * Update docs/modules/dqn.rst Co-authored-by: Antonin RAFFIN <antonin.raffin@ensta.org> * Added RMSProp in optimizer_kwargs, as described in nature paper * Exploration fraction is inverse of 50.000.000 (total frames) / 1.000.000 (frames with linear schedule) according to nature paper * Changelog update for buffer dtype * standard exlude parameters should be always excluded to assure proper saving only if intentionally included by ``include`` parameter * slightly more iterations on test_discrete to pass the test * added param use_rms_prop instead of mutable default argument * forgot alpha * using huber loss, adam and learning rate 1e-4 * account for train_freq in update_target_network * Added memory check for both buffers * Doc updated for buffer allocation * Added psutil Requirement * Adapted test_identity.py * Fixes with new SB3 version * Fix for tensorboard name * Convert assert to warning and fix tests * Refactor off-policy algorithms * Fixes * test: remove next_obs in replay buffer * Update changelog * Fix tests and use tmp_path where possible * Fix sampling bug in buffer * Do not store next obs on episode termination * Fix replay buffer sampling * Update comment * moved epsilon from policy to model * Update predict method * Update atari wrappers to match SB2 * Minor edit in the buffers * Update changelog * Merge branch 'master' into dqn * Update DQN to new structure * Fix tests and remove hardcoded path * Fix for DQN * Disable memory efficient replay buffer by default * Fix docstring * Add tests for memory efficient buffer * Update changelog * Split collect rollout * Move target update outside `train()` for DQN * Update changelog * Update linear schedule doc * Cleanup DQN code * Minor edit * Update version and docker images Co-authored-by: Antonin RAFFIN <antonin.raffin@ensta.org>
145 lines
5.5 KiB
Python
145 lines
5.5 KiB
Python
import os
|
|
import shutil
|
|
|
|
import pytest
|
|
import gym
|
|
import numpy as np
|
|
|
|
from stable_baselines3 import A2C
|
|
from stable_baselines3.common.monitor import Monitor
|
|
from stable_baselines3.common.atari_wrappers import ClipRewardEnv
|
|
from stable_baselines3.common.evaluation import evaluate_policy
|
|
from stable_baselines3.common.cmd_util import make_vec_env, make_atari_env
|
|
from stable_baselines3.common.vec_env import DummyVecEnv, SubprocVecEnv
|
|
from stable_baselines3.common.noise import (VectorizedActionNoise,
|
|
OrnsteinUhlenbeckActionNoise, ActionNoise)
|
|
|
|
|
|
@pytest.mark.parametrize("env_id", ['CartPole-v1', lambda: gym.make('CartPole-v1')])
|
|
@pytest.mark.parametrize("n_envs", [1, 2])
|
|
@pytest.mark.parametrize("vec_env_cls", [None, SubprocVecEnv])
|
|
@pytest.mark.parametrize("wrapper_class", [None, gym.wrappers.TimeLimit])
|
|
def test_make_vec_env(env_id, n_envs, vec_env_cls, wrapper_class):
|
|
env = make_vec_env(env_id, n_envs, vec_env_cls=vec_env_cls,
|
|
wrapper_class=wrapper_class, monitor_dir=None, seed=0)
|
|
|
|
assert env.num_envs == n_envs
|
|
|
|
if vec_env_cls is None:
|
|
assert isinstance(env, DummyVecEnv)
|
|
if wrapper_class is not None:
|
|
assert isinstance(env.envs[0], wrapper_class)
|
|
else:
|
|
assert isinstance(env.envs[0], Monitor)
|
|
else:
|
|
assert isinstance(env, SubprocVecEnv)
|
|
# Kill subprocesses
|
|
env.close()
|
|
|
|
|
|
@pytest.mark.parametrize("env_id", ['BreakoutNoFrameskip-v4'])
|
|
@pytest.mark.parametrize("n_envs", [1, 2])
|
|
@pytest.mark.parametrize("wrapper_kwargs", [None, dict(clip_reward=False, screen_size=60)])
|
|
def test_make_atari_env(env_id, n_envs, wrapper_kwargs):
|
|
env_id = 'BreakoutNoFrameskip-v4'
|
|
env = make_atari_env(env_id, n_envs,
|
|
wrapper_kwargs=wrapper_kwargs, monitor_dir=None, seed=0)
|
|
|
|
assert env.num_envs == n_envs
|
|
|
|
obs = env.reset()
|
|
|
|
new_obs, reward, _, _ = env.step([env.action_space.sample() for _ in range(n_envs)])
|
|
|
|
assert obs.shape == new_obs.shape
|
|
|
|
# Wrapped into DummyVecEnv
|
|
wrapped_atari_env = env.envs[0]
|
|
if wrapper_kwargs is not None:
|
|
assert obs.shape == (n_envs, 60, 60, 1)
|
|
assert wrapped_atari_env.observation_space.shape == (60, 60, 1)
|
|
assert not isinstance(wrapped_atari_env.env, ClipRewardEnv)
|
|
else:
|
|
assert obs.shape == (n_envs, 84, 84, 1)
|
|
assert wrapped_atari_env.observation_space.shape == (84, 84, 1)
|
|
assert isinstance(wrapped_atari_env.env, ClipRewardEnv)
|
|
assert np.max(np.abs(reward)) < 1.0
|
|
|
|
|
|
def test_custom_vec_env(tmp_path):
|
|
"""
|
|
Stand alone test for a special case (passing a custom VecEnv class) to avoid doubling the number of tests.
|
|
"""
|
|
monitor_dir = tmp_path / 'test_make_vec_env/'
|
|
env = make_vec_env('CartPole-v1', n_envs=1,
|
|
monitor_dir=monitor_dir, seed=0,
|
|
vec_env_cls=SubprocVecEnv, vec_env_kwargs={'start_method': None})
|
|
|
|
assert env.num_envs == 1
|
|
assert isinstance(env, SubprocVecEnv)
|
|
assert os.path.isdir(monitor_dir)
|
|
# Kill subprocess
|
|
env.close()
|
|
# Cleanup folder
|
|
shutil.rmtree(monitor_dir)
|
|
|
|
# This should fail because DummyVecEnv does not have any keyword argument
|
|
with pytest.raises(TypeError):
|
|
make_vec_env('CartPole-v1', n_envs=1, vec_env_kwargs={'dummy': False})
|
|
|
|
|
|
def test_evaluate_policy():
|
|
model = A2C('MlpPolicy', 'Pendulum-v0', seed=0)
|
|
n_steps_per_episode, n_eval_episodes = 200, 2
|
|
model.n_callback_calls = 0
|
|
|
|
def dummy_callback(locals_, _globals):
|
|
locals_['model'].n_callback_calls += 1
|
|
|
|
_, episode_lengths = evaluate_policy(model, model.get_env(), n_eval_episodes, deterministic=True,
|
|
render=False, callback=dummy_callback, reward_threshold=None,
|
|
return_episode_rewards=True)
|
|
|
|
n_steps = sum(episode_lengths)
|
|
assert n_steps == n_steps_per_episode * n_eval_episodes
|
|
assert n_steps == model.n_callback_calls
|
|
|
|
# Reaching a mean reward of zero is impossible with the Pendulum env
|
|
with pytest.raises(AssertionError):
|
|
evaluate_policy(model, model.get_env(), n_eval_episodes, reward_threshold=0.0)
|
|
|
|
episode_rewards, _ = evaluate_policy(model, model.get_env(), n_eval_episodes, return_episode_rewards=True)
|
|
assert len(episode_rewards) == n_eval_episodes
|
|
|
|
|
|
def test_vec_noise():
|
|
num_envs = 4
|
|
num_actions = 10
|
|
mu = np.zeros(num_actions)
|
|
sigma = np.ones(num_actions) * 0.4
|
|
base: ActionNoise = OrnsteinUhlenbeckActionNoise(mu, sigma)
|
|
with pytest.raises(ValueError):
|
|
vec = VectorizedActionNoise(base, -1)
|
|
with pytest.raises(ValueError):
|
|
vec = VectorizedActionNoise(base, None)
|
|
with pytest.raises(ValueError):
|
|
vec = VectorizedActionNoise(base, "whatever")
|
|
|
|
vec = VectorizedActionNoise(base, num_envs)
|
|
assert vec.n_envs == num_envs
|
|
assert vec().shape == (num_envs, num_actions)
|
|
assert not (vec() == base()).all()
|
|
with pytest.raises(ValueError):
|
|
vec = VectorizedActionNoise(None, num_envs)
|
|
with pytest.raises(TypeError):
|
|
vec = VectorizedActionNoise(12, num_envs)
|
|
with pytest.raises(AssertionError):
|
|
vec.noises = []
|
|
with pytest.raises(TypeError):
|
|
vec.noises = None
|
|
with pytest.raises(ValueError):
|
|
vec.noises = [None] * vec.n_envs
|
|
with pytest.raises(AssertionError):
|
|
vec.noises = [base] * (num_envs - 1)
|
|
assert all(isinstance(noise, type(base)) for noise in vec.noises)
|
|
assert len(vec.noises) == num_envs
|