An experimentation framework for Reinforcement Learning using OpenAI Gym, Tensorflow, and Keras.
MIT License
An experimentation framework for Reinforcement Learning using OpenAI Gym, Tensorflow, and Keras.
OpenAI Lab is created to do Reinforcement Learning (RL) like science - theorize, experiment. It provides an easy interface to OpenAI Gym and Keras, with an automated experimentation and evaluation framework.
With OpenAI Lab, we could focus on researching the essential elements of reinforcement learning such as the algorithm, policy, memory, and parameter tuning. It allows us to build agents efficiently using existing components with the implementations from research ideas. We could then test the research hypotheses systematically by running experiments.
Read more about the research problems the Lab addresses in Motivations. Ultimately, the Lab is a generalized framework for doing reinforcement learning, agnostic of OpenAI Gym and Keras. E.g. Pytorch-based implementations are on the roadmap.
A list of the core RL algorithms implemented/planned.
To see their scores against OpenAI gym environments, go to Fitness Matrix.
algorithm | implementation | eval score (pending) |
---|---|---|
DQN | DQN | - |
Double DQN | DoubleDQN | - |
Dueling DQN | - | - |
Sarsa | DeepSarsa | - |
Off-Policy Sarsa | OffPolicySarsa | - |
PER (Prioritized Experience Replay) | PrioritizedExperienceReplay | - |
CEM (Cross Entropy Method) | next | - |
REINFORCE | - | - |
DPG (Deterministic Policy Gradient) off-policy actor-critic | ActorCritic | - |
DDPG (Deep-DPG) actor-critic with target networks | DDPG | - |
A3C (asynchronous advantage actor-critic) | - | - |
Dyna | next | - |
TRPO | - | - |
Q*(lambda) | - | - |
Retrace(lambda) | - | - |
Neural Episodic Control (NEC) | - | - |
EWC (Elastic Weight Consolidation) | - | - |
Next, see Installation and jump to Quickstart.
Timelapse of OpenAI Lab, solving CartPole-v0.