Skip to yearly menu bar Skip to main content


Poster

Evolving Reinforcement Learning Algorithms

John Co-Reyes · Yingjie Miao · Daiyi Peng · Esteban Real · Quoc V Le · Sergey Levine · Honglak Lee · Aleksandra Faust

Virtual

Keywords: [ genetic programming ] [ evolutionary algorithms ] [ meta-learning ] [ reinforcement learning ]


Abstract:

We propose a method for meta-learning reinforcement learning algorithms by searching over the space of computational graphs which compute the loss function for a value-based model-free RL agent to optimize. The learned algorithms are domain-agnostic and can generalize to new environments not seen during training. Our method can both learn from scratch and bootstrap off known existing algorithms, like DQN, enabling interpretable modifications which improve performance. Learning from scratch on simple classical control and gridworld tasks, our method rediscovers the temporal-difference (TD) algorithm. Bootstrapped from DQN, we highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games. The analysis of the learned algorithm behavior shows resemblance to recently proposed RL algorithms that address overestimation in value-based methods.

Chat is not available.