Stable Baselines is a set of improved implementations of reinforcement learning algorithms based on OpenAI Baselines.
You can read a detailed presentation of Stable Baselines in the Medium article.
These algorithms will make it easier for the research community and industry to replicate, refine, and identify new ideas, and will create good baselines to build projects on top of. We expect these tools will be used as a base around which new ideas can be added, and as a tool for comparing a new approach against existing ones. We also hope that the simplicity of these tools will allow beginners to experiment with a more advanced toolset, without being buried in implementation details.
Note: despite its simplicity of use, Stable Baselines (SB) assumes you have some knowledge about Reinforcement Learning (RL). You should not utilize this library without some practice. To that extent, we provide good resources in the documentation to get started with RL.
This toolset is a fork of OpenAI Baselines, with a major structural refactoring, and code cleanups:
- Unified structure for all algorithms
- PEP8 compliant (unified code style)
- Documented functions and classes
- More tests & more code coverage
- Additional algorithms: SAC and TD3 (+ HER support for DQN, DDPG, SAC and TD3)
Features | Stable-Baselines | OpenAI Baselines |
---|---|---|
State of the art RL methods | ✔️ (1) | ✔️ |
Documentation | ✔️ | ❌ |
Custom environments | ✔️ | ✔️ |
Custom policies | ✔️ | ➖ (2) |
Common interface | ✔️ | ➖ (3) |
Tensorboard support | ✔️ | ➖ (4) |
Ipython / Notebook friendly | ✔️ | ❌ |
PEP8 code style | ✔️ | ✔️ (5) |
Custom callback | ✔️ | ➖ (6) |
(1): Forked from previous version of OpenAI baselines, with now SAC and TD3 in addition
(2): Currently not available for DDPG, and only from the run script.
(3): Only via the run script.
(4): Rudimentary logging of training information (no loss nor graph).
(5): EDIT: you did it OpenAI! 🐱
(6): Passing a callback function is only available for DQN
Documentation is available online: https://stable-baselines.readthedocs.io/
RL Baselines Zoo. is a collection of pre-trained Reinforcement Learning agents using Stable-Baselines.
It also provides basic scripts for training, evaluating agents, tuning hyperparameters and recording videos.
Goals of this repository:
- Provide a simple interface to train and enjoy RL agents
- Benchmark the different Reinforcement Learning algorithms
- Provide tuned hyperparameters for each environment and RL algorithm
- Have fun with the trained agents!
Github repo: https://github.com/araffin/rl-baselines-zoo
Documentation: https://stable-baselines.readthedocs.io/en/master/guide/rl_zoo.html
Note: Stabe-Baselines supports Tensorflow versions from 1.8.0 to 1.14.0. Support for Tensorflow 2 API is planned.
Baselines requires python3 (>=3.5) with the development headers. You'll also need system packages CMake, OpenMPI and zlib. Those can be installed as follows
sudo apt-get update && sudo apt-get install cmake libopenmpi-dev python3-dev zlib1g-dev
Installation of system packages on Mac requires Homebrew. With Homebrew installed, run the following:
brew install cmake openmpi
To install stable-baselines on Windows, please look at the documentation.
Install the Stable Baselines package:
pip install stable-baselines[mpi]
This includes an optional dependency on MPI, enabling algorithms DDPG, GAIL, PPO1 and TRPO. If you do not need these algorithms, you can install without MPI:
pip install stable-baselines
Please read the documentation for more details and alternatives (from source, using docker).
Most of the library tries to follow a sklearn-like syntax for the Reinforcement Learning algorithms.
Here is a quick example of how to train and run PPO2 on a cartpole environment:
import gym
from stable_baselines.common.policies import MlpPolicy
from stable_baselines.common.vec_env import DummyVecEnv
from stable_baselines import PPO2
env = gym.make('CartPole-v1')
env = DummyVecEnv([lambda: env]) # The algorithms require a vectorized environment to run
model = PPO2(MlpPolicy, env, verbose=1)
model.learn(total_timesteps=10000)
obs = env.reset()
for i in range(1000):
action, _states = model.predict(obs)
obs, rewards, dones, info = env.step(action)
env.render()
env.close()
Or just train a model with a one liner if the environment is registered in Gym and if the policy is registered:
from stable_baselines import PPO2
model = PPO2('MlpPolicy', 'CartPole-v1').learn(10000)
Please read the documentation for more examples.
All the following examples can be executed online using Google colab notebooks:
Name | Refactored(1) | Recurrent | Box |
Discrete |
MultiDiscrete |
MultiBinary |
Multi Processing |
---|---|---|---|---|---|---|---|
A2C | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ |
ACER | ✔️ | ✔️ | ❌ (5) | ✔️ | ❌ | ❌ | ✔️ |
ACKTR | ✔️ | ✔️ | ✔️ | ✔️ | ❌ | ❌ | ✔️ |
DDPG | ✔️ | ❌ | ✔️ | ❌ | ❌ | ❌ | ✔️ (4) |
DQN | ✔️ | ❌ | ❌ | ✔️ | ❌ | ❌ | ❌ |
GAIL (2) | ✔️ | ❌ | ✔️ | ✔️ | ❌ | ❌ | ✔️ (4) |
HER (3) | ✔️ | ❌ | ✔️ | ✔️ | ❌ | ✔️ | ❌ |
PPO1 | ✔️ | ❌ | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ (4) |
PPO2 | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ |
SAC | ✔️ | ❌ | ✔️ | ❌ | ❌ | ❌ | ❌ |
TD3 | ✔️ | ❌ | ✔️ | ❌ | ❌ | ❌ | ❌ |
TRPO | ✔️ | ❌ | ✔️ | ✔️ | ✔️ | ✔️ | ✔️ (4) |
(1): Whether or not the algorithm has be refactored to fit the BaseRLModel
class.
(2): Only implemented for TRPO.
(3): Re-implemented from scratch, now supports DQN, DDPG, SAC and TD3
(4): Multi Processing with MPI.
(5): TODO, in project scope.
NOTE: Soft Actor-Critic (SAC) and Twin Delayed DDPG (TD3) were not part of the original baselines and HER was reimplemented from scratch.
Actions gym.spaces
:
Box
: A N-dimensional box that containes every point in the action space.Discrete
: A list of possible actions, where each timestep only one of the actions can be used.MultiDiscrete
: A list of possible actions, where each timestep only one action of each discrete set can be used.MultiBinary
: A list of possible actions, where each timestep any of the actions can be used in any combination.
Some of the baselines examples use MuJoCo (multi-joint dynamics in contact) physics simulator, which is proprietary and requires binaries and a license (temporary 30-day license can be obtained from www.mujoco.org). Instructions on setting up MuJoCo can be found here
All unit tests in baselines can be run using pytest runner:
pip install pytest pytest-cov
pytest --cov-config .coveragerc --cov-report html --cov-report term --cov=.
We try to maintain a list of project using stable-baselines in the documentation, please tell us when if you want your project to appear on this page ;)
To cite this repository in publications:
@misc{stable-baselines,
author = {Hill, Ashley and Raffin, Antonin and Ernestus, Maximilian and Gleave, Adam and Kanervisto, Anssi and Traore, Rene and Dhariwal, Prafulla and Hesse, Christopher and Klimov, Oleg and Nichol, Alex and Plappert, Matthias and Radford, Alec and Schulman, John and Sidor, Szymon and Wu, Yuhuai},
title = {Stable Baselines},
year = {2018},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/hill-a/stable-baselines}},
}
Stable-Baselines is currently maintained by Ashley Hill (aka @hill-a), Antonin Raffin (aka @araffin), Maximilian Ernestus (aka @erniejunior), Adam Gleave (@AdamGleave) and Anssi Kanervisto (@Miffyli).
Important Note: We do not do technical support, nor consulting and don't answer personal questions per email.
To any interested in making the baselines better, there is still some documentation that needs to be done. If you want to contribute, please read CONTRIBUTING.md guide first.
Stable Baselines was created in the robotics lab U2IS (INRIA Flowers team) at ENSTA ParisTech.
Logo credits: L.M. Tenkes