stable-baselines3-contrib-sacd/tests/test_deterministic.py

61 lines
2.2 KiB
Python

import pytest
from stable_baselines3.common.env_util import make_vec_env
from stable_baselines3.common.noise import NormalActionNoise
from stable_baselines3.common.vec_env import VecNormalize
from sb3_contrib import ARS, QRDQN, TQC, RecurrentPPO
from sb3_contrib.common.vec_env import AsyncEval
N_STEPS_TRAINING = 500
SEED = 0
ARS_MULTI = "ars_multi"
@pytest.mark.parametrize("algo", [ARS, QRDQN, TQC, ARS_MULTI, RecurrentPPO])
def test_deterministic_training_common(algo):
results = [[], []]
rewards = [[], []]
# Smaller network
kwargs = {"policy_kwargs": dict(net_arch=[64])}
env_id = "Pendulum-v1"
if algo == ARS_MULTI:
algo = ARS
ars_multi = True
else:
ars_multi = False
if algo in [TQC]:
kwargs.update({"action_noise": NormalActionNoise(0.0, 0.1), "learning_starts": 100, "train_freq": 4})
else:
if algo == QRDQN:
env_id = "CartPole-v1"
kwargs.update({"learning_starts": 100, "target_update_interval": 100})
elif algo == ARS:
kwargs.update({"n_delta": 2})
elif algo == RecurrentPPO:
kwargs.update({"policy_kwargs": dict(net_arch=[], enable_critic_lstm=True, lstm_hidden_size=8)})
kwargs.update({"n_steps": 50, "n_epochs": 4})
policy_str = "MlpLstmPolicy" if algo == RecurrentPPO else "MlpPolicy"
for i in range(2):
model = algo(policy_str, env_id, seed=SEED, **kwargs)
learn_kwargs = {"total_timesteps": N_STEPS_TRAINING}
if ars_multi:
learn_kwargs["async_eval"] = AsyncEval(
[lambda: VecNormalize(make_vec_env(env_id, seed=SEED, n_envs=1)) for _ in range(2)],
model.policy,
)
model.learn(**learn_kwargs)
env = model.get_env()
obs = env.reset()
states = None
episode_start = None
for _ in range(100):
action, states = model.predict(obs, state=states, episode_start=episode_start, deterministic=False)
obs, reward, episode_start, _ = env.step(action)
results[i].append(action)
rewards[i].append(reward)
assert sum(results[0]) == sum(results[1]), results
assert sum(rewards[0]) == sum(rewards[1]), rewards