Zeta36 / Asynchronous-Methods-for-Deep-Reinforcement-LearningView on GitHub
Using a paper from Google DeepMind I've developed a new version of the DQN using threads exploration instead of memory replay as explain in here: http://arxiv.org/pdf/1602.01783v1.pdf I used the one-step-Q-learning pseudocode, and now we can train the Pong game in less than 20 hours and without any GPU or network distribution.
84Mar 4, 2016Updated 10 years ago

Alternatives and similar repositories for Asynchronous-Methods-for-Deep-Reinforcement-Learning

Users that are interested in Asynchronous-Methods-for-Deep-Reinforcement-Learning are comparing it to the libraries listed below

Sorting:

Are these results useful?