import pytest from stable_baselines3.common.env_util import make_vec_env from stable_baselines3.common.noise import NormalActionNoise from stable_baselines3.common.vec_env import VecNormalize from sb3_contrib import ARS, QRDQN, TQC, RecurrentPPO from sb3_contrib.common.vec_env import AsyncEval N_STEPS_TRAINING = 500 SEED = 0 ARS_MULTI = "ars_multi" @pytest.mark.parametrize("algo", [ARS, QRDQN, TQC, ARS_MULTI, RecurrentPPO]) def test_deterministic_training_common(algo): results = [[], []] rewards = [[], []] # Smaller network kwargs = {"policy_kwargs": dict(net_arch=[64])} env_id = "Pendulum-v1" if algo == ARS_MULTI: algo = ARS ars_multi = True else: ars_multi = False if algo in [TQC]: kwargs.update({"action_noise": NormalActionNoise(0.0, 0.1), "learning_starts": 100, "train_freq": 4}) else: if algo == QRDQN: env_id = "CartPole-v1" kwargs.update({"learning_starts": 100, "target_update_interval": 100}) elif algo == ARS: kwargs.update({"n_delta": 2}) elif algo == RecurrentPPO: kwargs.update({"policy_kwargs": dict(net_arch=[], enable_critic_lstm=True, lstm_hidden_size=8)}) kwargs.update({"n_steps": 50, "n_epochs": 4}) policy_str = "MlpLstmPolicy" if algo == RecurrentPPO else "MlpPolicy" for i in range(2): model = algo(policy_str, env_id, seed=SEED, **kwargs) learn_kwargs = {"total_timesteps": N_STEPS_TRAINING} if ars_multi: learn_kwargs["async_eval"] = AsyncEval( [lambda: VecNormalize(make_vec_env(env_id, seed=SEED, n_envs=1)) for _ in range(2)], model.policy, ) model.learn(**learn_kwargs) env = model.get_env() obs = env.reset() states = None episode_start = None for _ in range(100): action, states = model.predict(obs, state=states, episode_start=episode_start, deterministic=False) obs, reward, episode_start, _ = env.step(action) results[i].append(action) rewards[i].append(reward) assert sum(results[0]) == sum(results[1]), results assert sum(rewards[0]) == sum(rewards[1]), rewards