Visual Intention Classification by Deep Learning for Gaze-based Human-Robot Interaction
Lei Shi, Cosmin Copot, Steve Vanlanduit
IFAC-PapersOnLine, 53(5), pp. 750-755, 2020.
Abstract
In this work, we propose a deep learning model to classify a human’s visual intention in gaze-based Human-Robot Interaction(HRI). We consider a scenario in which a human wears a pair of eye tracking glasses and can select an object by gaze and a robotic manipulator picks up the object. A neural network is trained as a binary classifier to classify if a human is looking at an object. The network architecture is based on Fully Convolutional Net(FCN), Convolutional Block Attention Modules(CBAM) and Residual Blocks. We evaluate our model with two experiments. In one experiment we test the performance in the scenario where only a single object exists and the other one multiple objects exist. The results show that our proposed network is accurate and it can generalize well. The F1 score on the single object is 0.971 and 0.962 on multiple objects.Links
BibTeX
@article{shi20_ifac,
author = {Shi, Lei and Copot, Cosmin and Vanlanduit, Steve},
title = {Visual Intention Classification by Deep Learning for Gaze-based Human-Robot Interaction},
journal = {IFAC-PapersOnLine},
year = {2020},
volume = {53},
number = {5},
pages = {750-755},
doi = {10.1016/j.ifacol.2021.04.168}
}