Skip to content

Latest commit

 

History

History
19 lines (19 loc) · 1.27 KB

README.md

File metadata and controls

19 lines (19 loc) · 1.27 KB

GoExplore-Atari-PyTorch

Implementation of First return, then explore (Go-Explore) by Adrien Ecoffet, Joost Huizinga, Joel Lehman, Kenneth O. Stanley, Jeff Clune. The result is a neural network policy that reaches a score of 2500 on the Atari environment MontezumaRevenge.

Content

  1. Exploration Phase with demonstration generation
  2. Robustification Phase (PPO + SIL + Backward algorithm)

Results

  1. Exploration Phase demos with 2500 score
    1 2 4 5

  2. Robustification Phase with 2500 score
    9

  3. Robustification Phase backward algorithm progress
    10

  4. Robustification Phase path
    11