Beyond DQN/A3C: A Survey in Advanced Reinforcement Learning
One of my favorite things about deep reinforcement learning is that, unlike supervised learning, it really, really doesn't want to work. Throwing a neural net at a computer vision problem might get you 80% of the way there. Throwing a neural net at an RL problem will probably blow something up in front of your face -- and it will blow up in a different way each time you try. A lot of the biggest challenges in RL revolve around two questions: how we interact with the environment effectively (e.g. In this post, I want to explore a few recent directions in deep RL research that attempt to address these challenges, and do so with particularly elegant parallels to human cognition. This post will begin with a quick review of two canonical deep RL algorithms -- DQN and A3C -- to provide us some intuitions to refer back to, and then jump into a deep dive on a few recent papers and breakthroughs in the categories described above.
Nov-12-2019, 23:06:32 GMT
- Country:
- Europe > United Kingdom > England > Greater London > London > Camden (0.04)
- Industry:
- Health & Medicine > Therapeutic Area (0.48)
- Technology: