Loading…

Neural network based reinforcement learning for audio–visual gaze control in human–robot interaction

•We address the problem of robot gaze control in human–robot social interactions.•The robot learns without human supervision, via a reinforcement learning approach.•Our reinforcement learning method uses neural networks on multimodal data.•Pretraining on a simulated environment avoids long training...

Full description

Saved in:
Bibliographic Details
Published in:Pattern recognition letters 2019-02, Vol.118, p.61-71
Main Authors: Lathuilière, Stéphane, Massé, Benoit, Mesejo, Pablo, Horaud, Radu
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:•We address the problem of robot gaze control in human–robot social interactions.•The robot learns without human supervision, via a reinforcement learning approach.•Our reinforcement learning method uses neural networks on multimodal data.•Pretraining on a simulated environment avoids long training periods with humans.•We test our proposal on simulated data, and on offline and online real data. [Display omitted] This paper introduces a novel neural network-based reinforcement learning approach for robot gaze control. Our approach enables a robot to learn and to adapt its gaze control strategy for human–robot interaction neither with the use of external sensors nor with human supervision. The robot learns to focus its attention onto groups of people from its own audio-visual experiences, independently of the number of people, of their positions and of their physical appearances. In particular, we use a recurrent neural network architecture in combination with Q-learning to find an optimal action-selection policy; we pre-train the network using a simulated environment that mimics realistic scenarios that involve speaking/silent participants, thus avoiding the need of tedious sessions of a robot interacting with people. Our experimental evaluation suggests that the proposed method is robust in terms of parameter configuration, i.e. the selection of the parameter values employed by the method do not have a decisive impact on the performance. The best results are obtained when both audio and visual information is jointly used. Experiments with the Nao robot indicate that our framework is a step forward towards the autonomous learning of a socially acceptable gaze behavior.
ISSN:0167-8655
1872-7344
DOI:10.1016/j.patrec.2018.05.023