Skip to yearly menu bar Skip to main content


Virtual presentation / poster accept

On the Data-Efficiency with Contrastive Image Transformation in Reinforcement Learning

Sicong Liu · Xi Zhang · Yushuo Li · Yifan Zhang · Jian Cheng

Keywords: [ Reinforcement Learning ] [ reinforcement learning ] [ self-supervised learning ] [ data augmentation ] [ representation learning ]


Abstract:

Data-efficiency has always been an essential issue in pixel-based reinforcement learning (RL). As the agent not only learns decision-making but also meaningful representations from images. The line of reinforcement learning with data augmentation shows significant improvements in sample-efficiency. However, it is challenging to guarantee the optimality invariant transformation, that is, the augmented data are readily recognized as a completely different state by the agent. In the end, we propose a contrastive invariant transformation (CoIT), a simple yet promising learnable data augmentation combined with standard model-free algorithms to improve sample-efficiency. Concretely, the differentiable CoIT leverages original samples with augmented samples and hastens the state encoder for a contrastive invariant embedding. We evaluate our approach on DeepMind Control Suite and Atari100K. Empirical results verify advances using CoIT, enabling it to outperform the new state-of-the-art on various tasks. Source code is available at https://github.com/mooricAnna/CoIT.

Chat is not available.