FOR.ai Reinforcement Learning Codebase 
 
 
 
Modular codebase for reinforcement learning models training, testing and visualization.
Contributors: Bryan M. Li, Alexander Cowen-Rivers, Piotr Kozakowski, David Tao, Siddhartha Rao Kamalakara, Nitarshan Rajkumar, Hariharan Sezhiyan, Sicong Huang, Aidan N. Gomez
- Agents: DQN, Vanilla Policy Gradient, DDPG, PPO
 - Environments:
 - Model-free asynchronous training  (
--num_workers) - Memory replay: Simple, Proportional Prioritized Experience Replay
 - Modularized
 
Example for recorded envrionment on various RL agents.
| MountainCar-v0 | Pendulum-v0 | VideoPinball-v0 | procgen-coinrun-v0 | 
|---|---|---|---|
![]()  | 
![]()  | 
![]()  | 
![]()  | 
It is recommended to install the codebase in a virtual environment (virtualenv or conda).
Configure use_gpu and (if on OSX) mac_package_manager (either macports or homebrew) params in setup.sh, then run it as
sh setup.shYou need to install the following for your system:
- TensorFlow
 - OpenAI Gym
 - OpenAI Atari
 - OpenAI ProcGen
 - FFmpeg
 - Additional python packages 
pip install -r ../requirements.txt 
# start training
python train.py --sys ... --hparams ... --output_dir ...
# run tensorboard
tensorboard --logdir ...
# test agnet
python train.py --sys ... --hparams ... --output_dir ... --test_only --render
Check available flags with --help, defaults.py for default hyper-parameters, and check hparams/dqn.py agent specific hyper-parameters examples.
hparams: Which hparams to use, defined under rl/hparamssys: Which system environment to use.env: Which RL environment to use.output_dir: The directory for model checkpoints and TensorBoard summary.train_steps:, Number of steps to train the agent.test_episodes: Number of episodes to test the agent.eval_episodes: Number of episodes to evaluate the agent.test_only: Test agent without training.copies: Number of independent training/testing runs to do.render: Render game play.record_video: Record game play.num_workers, number of workers.
More detailed documentation can be found here.
We'd love to accept your contributions to this project. Please feel free to open an issue, or submit a pull request as necessary. Contact us [email protected] for potential collaborations and joining FOR.ai.



