Skip to content

Latest commit

 

History

History
40 lines (25 loc) · 2.17 KB

README.md

File metadata and controls

40 lines (25 loc) · 2.17 KB

Deep Reinforcement Learning Codes

Currently, there are only the codes for distributional reinforcement learning here.

The codes for C51, QR-DQN, and IQN are a slight change from sungyubkim.

QUOTA is implemented based on the work of the algorithm's author: Shangtong Zhang.

Always up for a chat -- shoot me an email (kun_chu@outlook.com) if you'd like to discuss anything.

Dependency:

  • pytorch(>=1.0.0)
  • gym(=0.10.9)
  • numpy
  • matplotlib

Usage:

In order to run my code, you need to create two subdirectories under the main directory: ./data/model/ & ./data/plots/. These two directories are used to store the data.

When your computer's python environment satisfies the above dependencies, you can run the code. For example, enter:

python 3_iqn.py Breakout 

on the command line to run the algorithms in the Atari environment. You can change some specific parameters for the algorithms inside the codes.

After training, you can plot the results by running result_show.py with appropriate parameters.

References:

  1. Human-level control through deep reinforcement learning (DQN) [Paper] [Code]

  2. A Distributional Perspective on Reinforcement Learning (C51) [Paper] [Code]

  3. Distributional Reinforcement Learning with Quantile Regression (QR-DQN) [Paper] [Code]

  4. Implicit Quantile Networks for Distributional Reinforcement Learning (IQN) [Paper] [Code]

  5. QUOTA: The Quantile Option Architecture for Reinforcement Learning (QUOTA) [Paper] [Code]