-
Notifications
You must be signed in to change notification settings - Fork 261
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
General discussion on State-of-the-Art Research #82
Comments
CURIOSITY-DRIVEN LEARNING – EXPLORATION BY RANDOM NETWORK DISTILLATION OpenAI have recently published a paper describing a new architecture extension for dealing with 'hard exploration' problem in Atari games. By highly rewarding the policy on exploration of 'state with interest' that would be normally ignored due to there complexity
for better overview of the papar this blog offers also a nice diagram of the network. On the same topic Uber had announced on their blog that they have achieved a significantly better results on the 'hard exploration' problem. |
@Kismuz , it looks like Uber's new Go-Explore algorithm had some break-through |
Population Based Training (PBT) of Neural Networks DeepMind had published a paper last year for 'lazy' hyperparameter tuning by self discovery of optimal hyperparameter set. Each worker is working with a small permutation of hyperparameters and during training the framework evaluate best performing worker/s and change the other workers accordingly to keep exploring the optimal set. (algo was tested on A3C)
An implementation of PBT can be found in Ray - Tune library |
Unsupervised Predictive Memory in a Goal-Directed Agent DeepMind recently published a paper on which they have presented a new external memory architecture (MERLIN) that is based on research from neuroscience.
|
Soft actor-critic (SAC) algorithm form UC Berkeley and Google Brain:Blog post: https://bair.berkeley.edu/blog/2018/12/14/sac/ |
Wow this is really impressive, this algo have some very nice proprieties in addition to showing great results. |
@JacobHanouna, yes in general, not at the moment - Of course I do welcome any contribution regarding all the aspects mentioned. |
Recent high-level review from JPMorgan research group: |
not state of the art per se but interesting blog one of the papers in the blog is also interesting |
While learning a bit about Meta-learning I came across the topic of Deep Neuroevolution which belong to the field of Genetic Algorithms. Paper Repro: Deep Neuroevolution Using Evolutionary AutoML to Discover Neural Network Architectures |
Google/Deepmind's new paper "Learning Latent Dynamics for Planning from Pixels"
|
A lot of research in the field of RL is being done now days.
I thought it can be both interesting and productive to have a post that would bring new research from time to time that might be relevant to this project.
The text was updated successfully, but these errors were encountered: