This is an tensorflow implementation of the paper "Continuous control with deep reinforcement learning". The code is adapted from here with some improvement. The main difference between the MorvanZhou's version and this code is
- Using
tf.train.ExponentialMovingAverage
to soft update the target network instead oftf.assign
; - Changing the activation function
tf.nn.relu
totf.nn.elu
; - Using
tf.train.AdagradOptimizer
to train the neural network; - Batch Updating the parameters using accumulating averaged gradients instead of the mean of mini batch.
The speed is 4.0X faster than the MorvanZhou's version when running on the same desktop computer with GTX 960 GPU, and achieving more stable learning curve.
Typing python Train.py
in the terminal to run, make sure you've already installed tensorflow and open ai gym.
Any comment or suggestion is welcome, you can open an issue or contact me via "fengredrum@gmail.com".
More deep RL algorithm is coming soon!
- Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., ... & Wierstra, D. (2015). Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971.
- https://github.com/MorvanZhou/Reinforcement-learning-with-tensorflow
- https://github.com/songrotek/DDPG
- https://github.com/RuiShu/micro-projects/tree/master/tf-ema