This repository provides and in-depth analysis of several Deep Q-Network (DQN) approaches in terms of performance and hardware footprint. Our target mission is a partially observable navigation problem, modeled in Minecraft, a state-of-the-art training and testing environment for research focusing on lifelong learning. The aim of this work is to compare several approaches fairly on a common task.
- Python 3.5
- TensorFlow
- Configure and install the Malmo Platform
- Install Gym-Minecraft
- Double Deep Q-Network
- Stacked Double Deep Q-Network
- Recurrent Double Deep Q-Network
- Dueling Double Deep Q-Network
- Stacked Dueling Double Deep Q-Network
- Recurrent Dueling Double Deep Q-Network
We would like to thank Clément Romac and Pierre Leroy for their exploratory work on partially observable missions. Tambet Matiisen for his implementation of a flexible and reliable training and testing environment for Minecraft.
- Double DQNs:
- Recurrent DQNs:
- Dueling DQNs:
- Practical Approach to RL: