Skip to content

Exploring whether DRQN + action prior + state-based expert + history-based entropy-reduction expert

License

Notifications You must be signed in to change notification settings

zhihanyang2022/drqn

Repository files navigation

Modified from https://github.com/g6ling/Reinforcement-Learning-Pytorch-Cartpole.

How to use this repo:

  • Make sure that you have installed miniconda (for Linux, see https://docs.conda.io/projects/conda/en/latest/user-guide/install/linux.html)
  • cd into this repo
  • cd packages
  • conda create --name pomdpr python=3.7 where pomdpr stands for POMDP and Robotics
  • conda activate pomdpr
  • chmod +x install_packages.sh where chmod makes the bash script executable on your device
  • ./install_packages.sh installs numpy scipy torch gym PyYAML wandb and rl_parsers and gym-pomdp (these two are stored inside drqn/packages)
  • cd .. back to the top level of the repo
  • Test your installation using python algorithms/POMDP/3-DRQN-Store-State-HeavenHell/train.py --lr=0.00005 --use_experts=0 --seed=1 --debug_mode=1 where debug_mode=1 makes sure that wandb is not used
  • wandb login
  • Do anything you want now.

Note that if you modify the Heaven-Hell pomdp file (e.g., modify the initial belief or the starting state distribution) you will need to re-install gym-pomdp for the change to take effect.


Below is README from the original repo.

PyTorch CartPole Example

Simple Cartpole example writed with pytorch.

Why Cartpole?

Cartpole is very easy problem and is converged very fast in many case. So you can run this example in your computer(maybe it take just only 1~2 minitue).

Rainbow

  • DQN [1]
  • Double [2]
  • Duel [3]
  • Multi-step [4]
  • PER(Prioritized Experience Replay) [5]
  • Nosiy-Net [6]
  • Distributional(C51) [7]
  • Rainbow [8]

PG(Policy Gradient)

  • REINFORCE [9]
  • Actor Critic [10]
  • Advantage Actor Critic
  • GAE(Generalized Advantage Estimation) [12]
  • TNPG [20]
  • TRPO [13]
  • PPO - Single Version [14]

Parallel

Distributional DQN

Exploration

POMDP (With RNN)

  • DQN (use state stack)
  • DRQN [24] [25]
  • DRQN (use state stack)
  • DRQN (store Rnn State) [16]
  • R2D2 - Single Version [16]

Reference

[1]Playing Atari with Deep Reinforcement Learning
[2]Deep Reinforcement Learning with Double Q-learning
[3]Dueling Network Architectures for Deep Reinforcement Learning
[4]Reinforcement Learning: An Introduction
[5]Prioritized Experience Replay
[6]Noisy Networks for Exploration
[7]A Distributional Perspective on Reinforcement Learning
[8]Rainbow: Combining Improvements in Deep Reinforcement Learning
[9]Policy Gradient Methods for Reinforcement Learning with Function Approximation
[10]Actor-Critic Algorithms
[11]Asynchronous Methods for Deep Reinforcement Learning
[12]HIGH-DIMENSIONAL CONTINUOUS CONTROL USING GENERALIZED ADVANTAGE ESTIMATION
[13]Trust Region Policy Optimization
[14]Proximal Policy Optimization
[15]DISTRIBUTED PRIORITIZED EXPERIENCE REPLAY
[16]RECURRENT EXPERIENCE REPLAY IN DISTRIBUTED REINFORCEMENT LEARNING
[17]EXPLORATION BY RANDOM NETWORK DISTILLATION
[18]Distributional Reinforcement Learning with Quantile Regression
[19]Implicit Quantile Networks for Distributional Reinforcement Learning
[20]A Natural Policy Gradient
[21]SAMPLE EFFICIENT ACTOR-CRITIC WITH EXPERIENCE REPLAY
[22]Curiosity-driven Exploration by Self-supervised Prediction
[23]IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
[24]Deep Recurrent Q-Learning for Partially Observable MDPs
[25]Playing FPS Games with Deep Reinforcement Learning

Acknowledgements

Use Cuda

check this issue. g6ling/Reinforcement-Learning-Pytorch-Cartpole#1

About

Exploring whether DRQN + action prior + state-based expert + history-based entropy-reduction expert

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published