Reinforcement Learning Zoo for Simple Usage¶

RLzoo is a collection of the most practical reinforcement learning algorithms, frameworks and applications. It is implemented with Tensorflow 2.0 and API of neural network layers in TensorLayer 2, to provide a hands-on fast-developing approach for reinforcement learning practices and benchmarks. It supports basic toy-tests like OpenAI Gym and DeepMind Control Suite with very simple configurations. Moreover, RLzoo supports robot learning benchmark environment RLBench based on Vrep/Pyrep simulator. Other large-scale distributed training framework for more realistic scenarios with Unity 3D, Mujoco, Bullet Physics, etc, will be supported in the future. A Springer textbook is also provided, you can get the free PDF if your institute has Springer license.
Installation¶
Direct installation:
1 | pip install rlzoo
|
Install from the source code on github:
1 2 3 | git clone https://github.com/tensorlayer/RLzoo.git
cd RLzoo
pip install .
|
Quick Start¶
Open ./run_rlzoo.py
:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | from rlzoo.common.env_wrappers import build_env
from rlzoo.common.utils import call_default_params
from rlzoo.algorithms import TD3
# choose an algorithm
AlgName = 'TD3'
# chose an environment
EnvName = 'Pendulum-v0'
# select a corresponding environment type
EnvType = 'classic_control'
# build an environment with wrappers
env = build_env(EnvName, EnvType)
# call default parameters for the algorithm and learning process
alg_params, learn_params = call_default_params(env, EnvType, AlgName)
# instantiate the algorithm
alg = eval(AlgName+'(**alg_params)')
# start the training
alg.learn(env=env, mode='train', render=False, **learn_params)
# test after training
alg.learn(env=env, mode='test', render=True, **learn_params)
|
Run the example:
python run_rlzoo.py
Common¶
DRL Book¶

- You can get the free PDF if your institute has Springer license.
Deep reinforcement learning (DRL) relies on the intersection of reinforcement learning (RL) and deep learning (DL). It has been able to solve a wide range of complex decision-making tasks that were previously out of reach for a machine and famously contributed to the success of AlphaGo. Furthermore, it opens up numerous new applications in domains such as healthcare, robotics, smart grids, and finance.
Divided into three main parts, this book provides a comprehensive and self-contained introduction to DRL. The first part introduces the foundations of DL, RL and widely used DRL methods and discusses their implementation. The second part covers selected DRL research topics, which are useful for those wanting to specialize in DRL research. To help readers gain a deep understanding of DRL and quickly apply the techniques in practice, the third part presents mass applications, such as the intelligent transportation system and learning to run, with detailed explanations.
The book is intended for computer science students, both undergraduate and postgraduate, who would like to learn DRL from scratch, practice its implementation, and explore the research topics. This book also appeals to engineers and practitioners who do not have strong machine learning background, but want to quickly understand how DRL works and use the techniques in their applications.
Editors¶
- Hao Dong - Peking University
- Zihan Ding - Princeton University
- Shanghang Zhang - University of California, Berkeley
Authors¶
- Hao Dong - Peking University
- Zihan Ding - Princeton University
- Shanghang Zhang - University of California, Berkeley
- Hang Yuan - Oxford University
- Hongming Zhang - Peking University
- Jingqing Zhang - Imperial College London
- Yanhua Huang - Xiaohongshu Technology Co.
- Tianyang Yu - Nanchang University
- Huaqing Zhang - Google
- Ruitong Huang - Borealis AI

DRL Tutorial¶

Different from RLzoo for simple usage with high-level APIs, the RL tutorial aims to make the reinforcement learning tutorial simple, transparent and straight-forward with low-level APIs, as this would not only benefits new learners of reinforcement learning, but also provide convenience for senior researchers to testify their new ideas quickly.

Contributing¶
This project is under active development, if you want to join the core team, feel free to contact Zihan Ding at ….