View a PDF of the paper titled ViSaRL: Visual Reinforcement Learning Guided by Human Saliency, by Anthony Liang and 2 other authors
Abstract:Training robots to perform complex control tasks from high-dimensional pixel input using reinforcement learning (RL) is sample-inefficient, because image observations are comprised primarily of task-irrelevant information. By contrast, humans are able to visually attend to task-relevant objects and areas. Based on this insight, we introduce Visual Saliency-Guided Reinforcement Learning (ViSaRL). Using ViSaRL to learn visual representations significantly improves the success rate, sample efficiency, and generalization of an RL agent on diverse tasks including DeepMind Control benchmark, robot manipulation in simulation and on a real robot. We present approaches for incorporating saliency into both CNN and Transformer-based encoders. We show that visual representations learned using ViSaRL are robust to various sources of visual perturbations including perceptual noise and scene variations. ViSaRL nearly doubles success rate on the real-robot tasks compared to the baseline which does not use saliency.
Submission history
From: Anthony Liang [view email]
[v1]
Sat, 16 Mar 2024 14:52:26 UTC (14,412 KB)
[v2]
Tue, 10 Sep 2024 07:04:02 UTC (14,412 KB)
Source link
lol