Step-by-step tutorials from DQN to Rainbow. Every tutorial contains both of theoretical backgrounds and object-oriented implementation. Just pick any topic in which you are interested, and learn and execute right away!
- DQN [NBViewer] [Colab]
- DoubleDQN [NBViewer] [Colab]
- PrioritizedExperienceReplay [NBViewer] [Colab]
- DuelingNet [NBViewer] [Colab]
- NoisyNet [NBViewer] [Colab]
- CategoricalDQN [NBViewer] [Colab]
- N-stepLearning [NBViewer] [Colab]
- Rainbow [NBViewer] [Colab]
This repository is tested on Anaconda virtual environment with python 3.6.1+
$ conda create -n rainbow_is_all_you_need python=3.6.1
$ conda activate rainbow_is_all_you_need
First, clone the repository.
git clone https://github.com/Curt-Park/rainbow-is-all-you-need.git
cd rainbow-is-all-you-need
Secondly, install packages required to execute the code. Just type:
make dep
- V. Mnih et al., "Human-level control through deep reinforcement learning." Nature, 518 (7540):529โ533, 2015.
- van Hasselt et al., "Deep Reinforcement Learning with Double Q-learning." arXiv preprint arXiv:1509.06461, 2015.
- T. Schaul et al., "Prioritized Experience Replay." arXiv preprint arXiv:1511.05952, 2015.
- Z. Wang et al., "Dueling Network Architectures for Deep Reinforcement Learning." arXiv preprint arXiv:1511.06581, 2015.
- M. Fortunato et al., "Noisy Networks for Exploration." arXiv preprint arXiv:1706.10295, 2017.
- M. G. Bellemare et al., "A Distributional Perspective on Reinforcement Learning." arXiv preprint arXiv:1707.06887, 2017.
- R. S. Sutton, "Learning to predict by the methods of temporal differences." Machine learning, 3(1):9โ44, 1988.
- M. Hessel et al., "Rainbow: Combining Improvements in Deep Reinforcement Learning." arXiv preprint arXiv:1710.02298, 2017.
Thanks goes to these wonderful people (emoji key):
Jinwoo Park (Curt) ๐ป ๐ | Kyunghwan Kim ๐ป |
This project follows the all-contributors specification. Contributions of any kind welcome!